A policy is a one-step plan depending on observations/state. The
(observation dependent) two-step plan is obtained from the policy by
incorporating the next observation and performing the associated action. A
mullti-step plan is then dependent on the multiple future observations. A
plan that does not depend on future observations is a special case of this,
and maybe is what AI planning does (but I don't know much about it).

On Mon, Mar 16, 2015 at 1:33 AM, Piaget Modeler via AGI <[email protected]>
wrote:

>
> Thanks for that feedback.  Much appreciated.
>
> Any other viewpoints?
>
> Anyone...
>
> Anyone...
>
> Bueller...
>
> Bueller...
>
> ~PM
>
> ------------------------------
> From: [email protected]
> To: [email protected]
> Subject: RE: [agi] Plans vs. Policies
> Date: Mon, 16 Mar 2015 03:13:46 +0200
>
>
> In brief then...
>
> The example model by itself is no silver bullet. It is suggested that it
> be seamlessly integrated with 2, other quantum-based methodologies known to
> me. Recently-verified research on the potential for the 3 integration has
> proved successful. Thus, it forms a 3-body approach, which is significant
> in its holistic value. All 3 the base methodologies operate as a platform
> of synthesis in entangled and/or disentangled states of thesis and
> antithesis. By viewing it in this way, one would notice that it potentially
> offers full support for the content of your ongoing discussions on coding?
>
> In most-normal form, the ontological model already contains all the
> inherent, system policies. This is an emerging by-product of the
> engineering process. It is robust in that it has 2 hierarchies of control
> (Criticality and Priority). There are various strategies for dealing with
> the hierarchy at any level of abstraction.
>
> The inherent policies are transcribed into simple sentences, exploiting
> all the implications of the possible, compound functions (adaptive
> associations - not dependencies). Except for components of the value
> "outcome" - never present in the existentially mature ontological model as
> submitted - all other components are potentially self recursive.
> Cursiveness would probably emerge at the reductionist level. In general,
> all contexts contain their own, particular level (graining) of policies.
> These could be tweaked in terms of coarseness, and thus perspective.  Scope
> of work could be managed into programs and projects, budgets, and so on.
>
> As such, the ontology presents as a superpattern, which is repeated
> consistently throughout any life cycle of any ontological event. Any
> triggering of the system would be treated as a state and an event. States
> and events would hold different values of the same, related data. Ambiguity
> becomes a non issue. This would provide elasticity to allow for dynamic
> routing within the policies of the hierarchy and scalable, parallel
> processing. A basic, optimization algorithm would come in very handy at
> this point. Policy statements may further be converted into plans, but
> these may never be in conflict with the superposition policy framework in
> the "superpattern". Thus, plans become objectively testable for ontological
> viability.
>
> Once entangled, policies assume a watchdog role. However, due to the
> nature of the entangled hierachical standard, not all policies would have
> to consume system resources at all times. Thus, policies could be created
> and destroyed for functional value, without negatively affecting the
> e-governance competency and performance. During intital systems
> development, policies would become semantically embedded into processes,
> via rules. Processes would also include functions, procedural workflow
> (program logic), and data. The value of information (views, products, and
> reports) would depend on the event-eco-systemic and end-user requirements,
> but be determined via the collective elements of process.
>
> Overall, the ontology is supported by a problem-solving framework, which
> is fully integrated with the systems-management framework. The elements of
> the problem-solving framework are People, Organization, Technology and
> Process. These elements are synthesized by its own version of an inference
> engine, namely E-Governance. For practical purposes, all policies are
> registered in the e-Governance layer, but co-managed by the actual policies
> and a policy-inference component, and so on to reduction. This layer
> provides the external ontological boundary, or People seam, to all MMI
> events.
>
> Even further, the ontology is supported by a practical, systems-management
> framework, which is driven by various aspects of system events, across
> workflow (content), including closed and open-looped feedback (feedback),
> of the emergent value chain (competency) for each event of the
> instantiation of a plan. In addition, the ontology is directly strenghtened
> by a researched, systems model for generative methodology. By design, all
> parts of the whole speak the same systems language, and when in synthesis,
> should theoretically generate the exponential value of being worth more
> than the sum of its parts.
>
> Trade-offs? So far, field research has yielded only 1, significant
> tradeoff. This tradeoff occurs during the process of migrating a systems
> model (logical) to a functional process (logical). The absence of reliable
> science for process remains an actual constraint, but some strategies for
> testing the consistency of process has been designed to deal with possible
> quality and integrity issues which may arise during the migration.
>
> Further, for AGI, I would imagine tradoffs occurring in the form of
> programming language(s), frameworks, and the selection of networked
> platform(s). In short then, one potential process tradeoff and x number of
> emerging technical tradeoffs. These tradeoffs should be dealt with as
> constraints or system requirements within the various extended
> architectures. These could probably be included as a standard model in any
> API, or API bus. To determine its suitability to the ontology, all
> architectures have to be tested for retro compliance with the policy
> implications of the superpattern.
>
> In summary:
> Within this ontology, a Plans vs Policies construct cannot exist within
> the operational system. At a planning decision level, they could exist
> separately within the same spacetime, but still not in opposition. Thye
> remain semantically separated entities in their own right. This is where
> planning could be used in a staging role, without having any direct
> performance impact on the operational system, other than
> eventually-approved workload (semantically integrated) or demand (workflow
> management) on system resources.
>
> When finally implemented (institutionalized) all polices become events of
> plans. Whn implemented (stage gated) plans become events of programs and
> projects. Projects inherit policy-compliant processes, sub-plans, special
> instructions (as problems), functions, resources, time, costs, and
> semantics. It becomes the structure for all these elements of Plan in
> Action (as a solution mechanism or value chain). In purely machine terms,
> Project could be satisfied by a single character, status parameter of any
> element within the ontology. In psuedo code; If project status = m, set
> elements {a,b,c,d...} to 0, else 1. This would auto-generate a
> mutally-exclusive construct for that particular event instance (tweakable
> via a standards-driven range and/or threshhold of graining).
>
> Control over project elements may be strengthened by a cross-correlating
> state function, which would relatively affect the mutual-exclusivity
> construct in a dynamic manner if needs be, at optimal efficiency. For
> example, if any of the excluded elements should be assigned a particular
> 1/0 workflow status, a particular rule may update the construct
> automatically, and in theory exclude/invoke another element in one step,
> and/or destroy the memory-resident element if it were not required for any
> event-future processing. Thus, resource-utilization could be pruned on a
> real-time basis, advancing the overall platform and layered performance to
> a level of effective complexity.
>
> Thank you. This was useful to me to explain.
>
> I hope it was generally clear enough and pertinent to your questions.
>
>
> From: [email protected]
> To: [email protected]
> Subject: [agi] Plans vs. Policies
> Date: Sun, 15 Mar 2015 16:22:46 -0700
>
> Reinforcement Learning uses "policies" to select actions while most work
> in AI Planning emphasizes
> the construction and representation of a "plan" which consists of a
> sequence of actions (or a hierarchy
> of composite and primitive actions).  Kindly compare, contrast, evaluate
> trade-offs, and recommend
> the plans or policies approach
>
> Your rationale is appreciated.
>
> ~PM
>
>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/26941503-0abb15dc> |
> Modify <https://www.listbox.com/member/?&;> Your Subscription
> <http://www.listbox.com>
>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/19999924-4a978ccc> |
> Modify <https://www.listbox.com/member/?&;> Your Subscription
> <http://www.listbox.com>
>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/10872673-8f99760d> |
> Modify
> <https://www.listbox.com/member/?&;>
> Your Subscription <http://www.listbox.com>
>



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to