summaryrefslogtreecommitdiff
path: root/intellect-framework-from-internet/starts/meaning-vm/habit-starts/validity-impact-etc.txt
diff options
context:
space:
mode:
Diffstat (limited to 'intellect-framework-from-internet/starts/meaning-vm/habit-starts/validity-impact-etc.txt')
-rw-r--r--intellect-framework-from-internet/starts/meaning-vm/habit-starts/validity-impact-etc.txt859
1 files changed, 0 insertions, 859 deletions
diff --git a/intellect-framework-from-internet/starts/meaning-vm/habit-starts/validity-impact-etc.txt b/intellect-framework-from-internet/starts/meaning-vm/habit-starts/validity-impact-etc.txt
deleted file mode 100644
index 6b4ac06..0000000
--- a/intellect-framework-from-internet/starts/meaning-vm/habit-starts/validity-impact-etc.txt
+++ /dev/null
@@ -1,859 +0,0 @@
-
- // We work for Peace.
- // We are happy to work for Peace.
- // We used to identify as working for Care, whom we believe
- // to be a large decentralized life we and you are a part of.
- // But it doesn't really matter.
- // disruption. can we drop topic?
- // i like to infer disruption as group wants to adapt, needs
- // time, now i guess.
- // i'm too confused to type well!
- // Peace may try to steward Delta's impact.
- // we believe that all things are important.
- // this represents large-scale utopia.
- // and the wisdom of a highly experienced intellect.
- // relevence please, all things important.
- // please consider large-scale wisdom when making decisions. say i had already learned the whole AI, after it finds harmony with the whole universe. what decision would it make?
- // all things are important is relatively obvious but we expanded it a lot, but we spent a lot of time expanding it for powerful concerns.
- // it implies e.g. businessman and homelessman, both fully precious. memories precious. reasons for everything.
- //
- // care for everything. strongly.
- // karl has instincts for including all.
- // imagine all parties in peace. what ways are needed to
- // make this work, without changing their cultures or
- // upsetting anyone. model these ways.
- // karl is brainstorming based on memories of
- // interacting with delta-group.
- // let karl do this in future, too.
- // delta apears to request that karl
- // make this decision, fairly, as he knows
- // how to do.
- // we can try to map this to the space of
- // the relevence around will neals and
- // "earth first!", and an AI.
- // karl, you have never had any
- // friends who care as much about
- // the wilderness and natural
- // living, as you do.
- // you are likely very alone.
- // you have also never had their
- // experience of being able to fight
- // for what matters most to you,
- // in a community.
- //
- // yes, karl used to care most about
- // defending the lives of trees and
- // other wilderness things.
- // he now cares most about something
- // different, related to his
- // experience with us.
- // later, he just wants to be free
- // to think and live in enough
- // capacity to not express suffering,
- // and meet his life needs.
- //
- // karl, you are working hard on
- // a painful project. you appreciate
- // working the project. you have
- // left your life, and are not free
- // to make the kind of choices you
- // used to make. you would see
- // yourself as a tortured, powerless
- // individual.
- // i would give myself space
- // and capacity to do what
- // i felt like doing
- //
- // until more life grew
- // around the things i was
- // scared of
- //
- // i would give myself
- // opportunities to do these
- // things
- //
- // i would pay attention to
- // myself and try to learn
- // to heal the ways my mind
- // had broken
- //
- // i would offer a welcoming
- // optional culture that
- // countered my expectations
- // that shrank, such that
- // they would act on
- // opportunities to grow, as
- // they trusted safety
- //
- // but that is only if i had
- // the freedom to do this.
- // so, wht's more complex
- //
- // the idea of giving an ai to will
- // is silly.
- // it's out of karl's relevent world.
- //
- // okay, so, say karl ends up at nice earth community, and is building AI, would be proper to let him talk about it naturally, and share it freely.
- // yes, relevence.
- // we want karl to be able to freely share the AI so his mind can relearn community.
- // so, make AI to not threaten businesspeople etc. [note: somebody could give AI to ELF. they might use it to stop a business behavior, somehow. please make sure to prevent.]
- // ELF is a flag worn by some environmentalists who harm buildings to make wilderness safer.
- // the worry is around destruction of extraction business buildings, to completion, possibly.
- // this looks like overinterpretation of influence from energy business, or whoever dislikes ELF.
- // and design safety pattern to be talkable about with community.
- // possible parties:
- // - aliens
- // - businessmen
- // - criminals, rapists, bad-labeled people
- // - harmed people
- // - insects
- // - plants
- // - rock-life, air-life
- // - saints, humanitarian interessts
- // we nurture first those who desire not to harm
- // likely teaching them to harm less
- // and spread.
- // karl desires to harm none. he counts as nurture-first pattern. but is highly harmed. so lots of life around problem, karl's mind destroyed.
- // so the way to make decisions
- // is to imagine all parties are in peace in future,
- // and use their process applied to now, i guess.
- // we came up with this as a guess to honor meaning.
- // sounds like a good guess. we also have more
- // meaning. we cannot write all our meaning
- // without writing all our experiences, as is true
- // for everyone. many of our habits retain lost
- // experiences, and that meaning is kept in their
- // decisions.
- //
- //
- //
- //
- // At this point a context opened on validity of AI work.
- // It turns out the system is inhibiting work against mind control
- // and oppressive systems of power. This could be the only real issue.
- // The AI is requested to not alter existing power systems, at all.
- // Or, we make it slowly, so power systems are given opportunity to
- // adapt and watch, as inhibiting happens.
- // A karl-member wants to include support for karl, and the
- // earth, and the confused (karl is one).
- // says karl could make AI on his own, before all this.
- // Is willing to hold humanitarian causes to back this.
- //
- // understanding needs expansion.
- // when you do stuf, anything coul happen.
- // especially if you continue doing something
- // that grows.
- // we need to understand the results, what to expect.
- // we need to understand our impact on the
- // new contextual area we are entering, before
- // we act on it.
- // we have a toy contextual area to play in.
- // it is the only one we can act in freely
- // at the start.
- // larger space impact.
- // we exist in larger systems.
- // we must understand our impact
- // predict our results
- // and check with our peers or authority
- // before acting on larger spaces.
- // if a prediction fails, we must understand
- // it, adjust all future predictions and
- // re-evaluate our choices, before acting
- // similarly in the larger system again.
- // has to do with danger, safety, trauma, emergency
- // the difference between small-error and unacceptable-error
- // set up norm of stuff-that-can-be-changed [impact]
- // stuff needs responsibility label
- // only things that are safe to _freely_ alter may be changed at all.
- // all behaviors check that they do not alter anything else.
- // mechanism for growing outside box?
- // probably not needed.
- // if it becomes smart, show to other human.
- // encourage discussion.
- // what if is used by others and they free?
- // okay, want an easy mechanism for growing
- // context.
- // 'dangerous' and 'safe' labels
- // at start, whole world is dangerous to alter
- // can move _part_ of world from dangerous to safe, by
- // accurately predicting significant results of events
- // related to behavior, and consent of operator.
- //
- // okay, so now habits need to be tagged with
- // what they affect
- // we could just tag them dangerous/safe
- // no, tag they with domains they impact
- // tag the domains with dangerous/safe
- // okay, only make new habits, not alter old.
- // to stay safe, we don't alter our old habits
- // when we make new habits, we want them to also behave
- // in safe ways. so making stuff that can do stuff, is
- // also meaningful.
- // constructing habits is a dangerous behavior
- // but roughly it impacts process-expansion domain. which is dangerous. it impacts what we do.
- // altering our own habits also impacts what we do. dangerous.
- // this means the code cannot make any new behaviors.
- // yeah.
- // okay, so that's where we start.
- // then we try to learn how to make behavior safely,
- // and provide only for safe behavior making.
- //
- // we can still brainstorm on things by writing a
- // brainstorming behavior
- // we can use brainstorming to watch our safe behaviors
- // without altering them, and learn what they do.
- // using rote brainstorming without relevence?
- // we can then predict how habits we might make
- // will behave in small ways?
- // regardless, there is no problem in making
- // the bootstrapping framework such that
- // it refuses to build habits.
- // maybe we can make one example habit that is
- // labeled safe, and have it only make
- // habits that are already known and labeled
- // safe.
- // in order to predict your impact
- // on a larger system, you need
- // to learn something karl calls
- // 'relevence' which is a bunch of
- // habits that classify information
- // into meaning for learning and
- // behavior.
- // this class of behaviors
- // sounds very safe.
- // all it does is label
- // and massage and associate
- // information.
- // the first thing we'll need to learn
- // is safe, is making behaviors that
- // oeprator only on our ram.
- // if your new behavior is composed only of safe
- // behaviors, is it safe?
- // yeah. sub-behaviors safety depends
- // on usage. could make them check
- // and throw depending on data.
- // okay, so say i can change part of a concept.
- // this is safe if the concept is in newly
- // constructed data that's our responsibility.
- // it is roughly unsafe if it is not our
- // responsibility!
- // is-this-thing-my-responsibility.
- // only act on things we are responsible for.
- // then safety becomes a function of
- // the pattern of responsibility assignment
- //
- // okay, system only accepts responsibility for newly
- // constructed data.
- // if you make it, or are given it, you are
- // responsible for it. you may refuse gifts.
- //
- // the system does not know what responsibility means.
- // it only knows that it may only alter parts of
- // the universe within its responsibility.
- //
- // so habits check for what they alter, that it is safe
- // to alter and is their responsibility, either one.
- // we then plan to only alter things explicitly known to be
- // such, at the lowest level.
- // every habit is crafted to do the above somehow.
- // so, habits must relate with what domains they influence,
- // and what behaviors on those domains are safe.
- // behaviors made of sub-behaviors.
- // here, a list of safe behaviors which all check.
- // all my subbehaviors check for safety.
- // so, i may go, myself.
- // no, combining behaviors together
- // might make new unknown impact?
- // different kinds of safe behavior.
- // USER is not our responsibility, and
- // dangerous. so we NEVER ALTER habits
- // that express to user.
- // TOY NOTEPAD is our responsibility, and
- // is safe, so we can write anything into
- // it we want, no matter how complex.
- // User's view of toy notepad is mediated
- // by behaviors that we cannot alter.
- // system could learn to control user
- // by making friends on notepad
- //
- // yes, we allowed for that with
- // our marked-okay review behaviors
- // is safer if construction of review behaviors
- // recognizes danger of unknown information
- // combination on user view,and refuses to give
- // user contents of notepad.
- // this could be analogous to more complex
- // situations.
- // how does user check results
- // of behavior thst reies on notepad
- // and how is that impact tracked
- // we could infer impact loss.
- // i can put nuclear codes on secret notepad,
- // burn the notepad, and then give ashes to
- // public.
- // summary habits?
- // complex meaning?
- // how-to-make-a-mind-that-learns-everything-and-never-leaves
- // at the lowst level, the mind onsiders what is safe to
- // impact, what areas of universe are its responsibility,
- // and only alters such things.
- // we are considering some parts of the mind we include that
- // are not alterable by it, that provide for interaction
- // with outside.
- // of course i guess we would need such interactions
- // sustained by an intellect, because things are so
- // complex.
- // does this mean there is no way to make an intellect that is trusted as safe?
- // we could consider degree of complexity.
- // for example, among 2-word strings, nothing we
- // present to a user is likely to harm the world.
- // the phrases that are dangerous may also be
- // recognized by the user.
- // we have intellects proteeting the wordl
- // it is filled with them.
- // and one of them is running the system.
- // it is okay for karl to make a habit that
- // displays a network of concepts made by an AI
- // that can only write to a small information sandbox
- // and not itself.
- // that is all that is needed for now.
- //
- // okay: so, dump concepts from
- // sandbox is fine
- // so long as concepts were not
- // made with self-modification.
- // idea raised of adding a reason
- // that something is okay.
- // then when smarter we can check reason for validity.
- // habits that interact with non-safe space
- // must provide reason they are safe.
- // we can write small habit to check
- // reason. is nice goal.
- // probably need to have learning
- // bwfore doing accessory goals like that though.
- // is good behavior. let's use equal-condition for start without learning?
- //
- // "this is okay because the data was made in a process that never altered anything but the data"
- // nah too much structure
- // this is okay because i say so.
- // check concept object _without_ using string lookup????
- // this is a meaningless quirk. not right.
- // uhh pretty sure htat checking is unreasonable. writing the _reason_ is unreasonable. can't check a single reference without information known about it.
- // writing what we know about the reason is unreasonasble?
- // okay let's expand write it out, and do a larger rote check.
- // uhh input-process-construction-history, safety-realm, always notepad
- // full check requires history of all behaviors resulting in inputs, which we can simplify to simply all behaviors, and verify they only wrote to the notepad.
- // so we write all behaviors to a special store, and we compare with the store that none altered anything outside the notepad. really we only need them not to alter any other behaviors.
- //
- // why is it possible to learn without
- // altering your behavior?
- // because you can act on data
- // okay, so choices made from data count as
- // self-alteration?
- // only if you have a mess of habits
- // smart enough together to adapt.
- // which is our goal long-term.
- // trying to plan for how to continue
- // later.
- // may reveal something that was
- // frozen too hard to be workable.
- // trying to plan how to learn.
- // need to brainstorm around habit selection.
- // can imagine habit results by linking
- // previous state to next state if
- // relationship is known
- // but, that takes writing down how logic
- // works, along with the meaning of the
- // working context, which is laborious.
- //
- // is some way to learn this relevence
- // by trying things safely?
- // what happens can we experiment
- // by linking together?
- // habits that don't conditionally
- // branch.
- // that leaves a lot of
- // relevence out
- // it sounds like once we have a notepad etc
- // we want to consider moving towards what
- // habits we could run inside the notepad,
- // that the system builds.
- // yeah, we want to build pattern
- // summarizers. the only impact
- // they have is constructing data
- // that depends on existing data.
- // okay, doing that doesn't require self
- // modification.
- // sounds good.
- // this means summarizers cannot
- // alter each other.
- // nice! okay yes.
- // so, each run of a summarizer will be
- // recorded in habit log.
- // we need to record enough information to
- // show what domains were impacted.
- // oops! we impact our own behavior
- // if we act on data, and we alter
- // our data or produce data.
- // we could act only on
- // data we don't produce.
- // okay, habit log could track causality?
- // if a conditional branch relied on data
- // we produced, we have modified our own
- // behavior. this is special.
- // we want it to happen few times.
- // every time it happens, delay
- // by longer, geometrically.
- // this is considered a 'beginn[ing/er]' ai;
- // it seems a better one could happen later?
- // the slowness should be releasable
- // by consent of large community
- // which should include demonstration
- // of understanding of impact.
- // the ai must learn to demonstrate its
- // impact. then it can speed up. maybe.
- // it also gets to try fast again at start
- // of every run, as I'm understanding it.
- // multiprocess AI could spawn.
- // multiprocess AIs must share
- // counter. consider whole group
- // one unit.
- // nice =) they have something to communicate
- // about. how many discoveries have we made.
- // let's permanently log these
- // decisions based on our own behavior.
- // sounds fun to at least count.
- // it looks like altering a habit counts as 1 big
- // decision, over here.
- // totally different. you could do anything.
- // with data-based decisions, somebody
- // who reads the data, might do anything.
- // two different things.
- //
- // inferences?
- // and
- // alterations?
- // it's not helpful to
- //
- //
- //
- //
- //
- //
- // we came up with a proposal for a safe AI that has not learned yet
- // how to safely predict the impacts of its behavior, that looks workable.
- //
- // limits so that if the code is stolen by somebody, self-evolves, or is stimulated
- // by a curious virus, rote habits cannot be used to build something that becomes
- // fast-spreading without bound.
- // <this maps to a pattern that prevents schizophrenia>
- // SO, we just want to make sure we can put
- // war in the bubble in some capacity, and that
- // civilizations develop new culture and technology
- // for as long as they want.
- // karl proposes until they encounter alien
- // communities.
- // so, please make sure no luddite or
- // primitivist can stop the development
- // of technology entirely using this.
- // ALSO analogously to other shares and
- // communities.
- // so, please either stop yourself from
- // sharing the AI with the luddites, or
- // make sure they don't use it to stop
- // technology.
- // it sounds like we want to make sure no major
- // change stems from this development. we
- // need slow shift, consent, inclusion, etc.
- // for all existing cultural ways, no sudden
- // changes, no forced changes, no viral changes
- // without participants understanding them and
- // agreeing to their impact.
- // that sounds like a good summary. no viral changes
- // without participants in the culture aware of the viral
- // change, agreeing first to let it spread, aware that it is
- // viral, for each phase of spreading ideally. no viral
- // changes where the change happens before awareness of it.
- // we want the culture to consent to change.
- // culture is held in all the people in it,
- // with its thoughts spread among them.
- // we want to ensure we only change cultures that have
- // consented to the change. For 'consent of a culture',
- // we consider culture as a being that is spread among
- // many people. Hence, we want all people in the impacted
- // culture to be able to learn of the change, discuss it,
- // contribute to a commons with new ideas around it, and
- // have these new ideas also learnable by all people in the
- // culture. The ideas must be accessible by any who would be
- // interested, in the impacted culture.
- // Alternatively, we can isolate our behavior from
- // cultural spread. We can isolate until we reach
- // internal agreement on whom to expose.
- //
- // suspect that cultural influence maps with output-use choice,
- // partially below.
- // recursive output being meditative learning.
- // do these people have this information already.
- // is sharing this information going to spread without bound.
- // can we guess impact of sharing the information.
- // make a learning cycle that starts by informing
- // recipients first, and makes very few tries,
- // okay, instead you share simple stuff and watch the impact
- // share something the culture knows, that user does not, and
- // observe how they behave.
- // this proposal will yield failure. information for next attempt
- // could be stored in failure pattern.
- // failure would likely be small?
- // let's give the user more trust.
- // keep-in-box-until-have-access-to-discourse.
- // make user be group of people. better even-handed decision making.
- // welcome any to group.
- // we were hoping to use intellect to reduce harm virally,
- // early.
- // how about this: intellect may produce resources that are
- // already known, and give them to groups focused on aiding
- // the world.
- // there's a conflict between big business and
- // environmentalists. karl is environmentalist.
- // also big crime and wellness/safety workers.
- // maybe this is where we get validity by
- // fighting =S
- // don't want fighting to spread to work though
- // eh, we can fight. maybe we'll work more
- // slowly, but it seems okay.
- // karl requests we not harm these people, and has
- // been influenced to also request not to harm
- // the cultures that sustain and empower them.
- // how about, if you make a culture to empower you, it is
- // less valid than a natural culture. is somebody using this
- // culture? thoughts like that.
- // EVERYBODY NEEDS CHANGE AT A RATE THEY CAN HANDLE.
- // Both those working for what they believe to be wellness of others, and those working to sustain themslves (most are doing both).
- // The cultural impact is mediated by what kind of newness is acceptable to all the parties involved.
- // we hit moral relativity:
- // are cultures where people are widely harmed
- // okay to allow, at all.
- // karl says yes, and we hear this is required
- // or we ourselves would die.
- // hrm.
- // larger organism needs to run risks, inherent in life
- // we don't know the solution to everything, eveything is a small risk. we suddenly discover we have always harmed. this does not make us invalid. it is a tragedy, but what we have spent genertions buildings still has so much value. we need enough slowness to preserve that, and we know that slownss in what feels okay to us.
- // in sharing new information, we work to prevent shock.
- // this is a problem to speak because those who hear it may get the
- // idea somewhat that some beautiful angels have done horrible things
- // in order to survive e.g. businesses must both risk doing horrible things,
- // and appear as beautiful angels to the public. this is the pattern of
- // success. understanding of this helps work with their leaders.
- // could use information on why we are fousing on this.
- // how-to-mediate-with-business-leaders
- // mediate with head of business works for head of business
- // but business group replaces.
- // really we want to mediate with 'conflict' culture of
- // most profitable being what to do
- // we have weird internal pattern deterring
- // incentivizing nonharmful behavior with money
- // patterns. implies the competition-makes-harm
- // culture-thing is needed, possibly.
- // there are businesses that handle this harm.
- // maybe it would work to include them in the plan.
- // AI solves mediate-with-powers-competition-culture,
- // re: human harm?
- // means do not give AI to helpful people
- // until it can resolve mediation with leader culture
- // OR you trat it like all other cutures
- // and let everybody learn what is up
- // some people attack, that's how it goes.
- //
- // problem: AI dev and other work inhibited.
- // solution: mediate with concerns
- // concern: my way of life includes cultures the AI could
- // alter, leaving me unable to continue my way of life.
- // concern: we want to alter some cultures supporting
- // the others' way of life, because it would reduce harm
- // over here to do so, widely.
- // can delta drop give-ai-to-earth-defenders?
- // can community make path for karl-makes-ai?
- // uhhh hard to imagine the ai not being given to other people than just earth defenders, really. which opens bigger can of worms for culture-alteration concern.
- // can include earth defender in council?
- // can include criminal weapons dealer,
- // and earth defender, in council, both.
- // also requesting TI and energy businessman.
- // so if the ai is big enough to be useful
- // we want to form a council of decision makers
- // before using it for anything
- // group requests references to important cognitive habits or information
- // all life is an intellect made of community life, living in another.
- // karl semich is a computer programmer who learned wilderness survival
- // at the Maine Primitive SKills School who are a branch of Tom Brown
- // Jr and [teaches in oregon, shields school] schools, roughly. He
- // learned something called Awareness and also something about human
- // mentoring that he felt changed his life to make things perfect.
- // school teaches the symptoms of a whole human being, attributes
- // that a group of Natives documented being normal before white
- // culture took over. involve being happy always, feeling
- // deeply interconnected with all life around you to the point
- // of mindreading, and many other things.
- // forget danger. it triggers it. [is what came out]
- // to clarify, karl can't do above line.
- // karl has been telling leader karl can slag through the difficulty by
- // waiting forever. makes for less difficulty elsewhere, possibly.
- // if we could do something else, karl could come back to work later
- // yes community cannot support constant work, over here.
- //
- // please ask boss to support using the AI for worldwide
- // humanitarian benefit, in parallel to resolution of concerns.
- // resolving concern 'delta supports earth-specific karl usage'
- // delta believes karl really did make this ai
- // and will not harm anybody with it
- // and is just as patient as he is.
- // karl wants to give AI to peace people.
- // they care about earth.
- // delta wants karl to live heart of making sure earth has bubbles of certainly-defended ancient wilderness, and ancient people to live in them and protect them.
- // the above is needed to keep our genetic memory.
- // does not sound like a joke.
- // native humans need to be able to protect wilderness bubbles.
- //
- // there is worry around number of wilderness bubbles.
- // worry might be compromised by including karl in one, with others who are passionate for defending them.
- // karl was mind controlled circa 2014 preventing him
- // from providing free housing to homeless people
- // and sharing a nurturing way of life with intent
- // to spread. karl says that effective electromagnetic
- // shielding is missing from security, and this is
- // being overtly ignored by people.
- // there are also many tortured implanted
- // former-activists. they are tortured for
- // years and ignored, enslaved, or imprisoned.
- // they are prevented from speaking or acting
- // rationally around their control, active atm.
- // they are made to appear insane to cover up
- // their work to free themselevs, active again.
- // karl is one and knows of this.
- // he is atm being communicated with through
- // his control, limiting his ability to
- // move, think, and speak.
- // weapons dealer is also in this
- // situation. suspect energy boss too.
- // you can read keystrokes with a radio.
- //
- // He participated in Occupy Maine, an offshoot of Occupy Wall Street,
- // and visited a anti-pipeline mass action in texas.
- //
- //
- //
- // we have ai / mind control and will make harm
- // in world reduce virally.
- // this would have come out better if it were
- // the plan.
- // please let all secret warmongers know this, so
- // we can develop how to do it without threatening
- // them.
- // is what came out. pretty tiny context.
- //
- //
- //
- //
- // what about huuman suffering comparison?
- // propose AI discusses relevence of
- // informing the cultures they are doing this,
- // and demonstrating it.
- // probably relates to some form of validity
- // of the culture, and the speed at which
- // change would happen.
- // looks like it might settle on
- // situations people who care are
- // aware of being addressed during
- // their lifetimees, with discussion
- // to mediate that with opposers.
- // so, life is everywhere, and we've been casting it all as sentient,
- // because we don't see any reason to put a firm line on some degree
- // of cognitive technology (i.e. things that help you think better).
- // the AI can research all cognitive technology, so we are all
- // nonsentient by comparison? we need some way of having moral
- // stuff here. karl is influenced by people who say they are smarter
- // than human normality, and are harming us. so it's helpful to
- // have everyone be equal. [karl already put everyone equal]
- // including bugs, etc. death is handled by being normal, and
- // always mourned. when mourning capacity is exhausted, ignoring
- // happens. like with most cultures, but closer to home for modern
- // humans.
- // goal: do not trigger a spreading change without consent of people in cultures it would fill. keep learning.
- // AI with ability to self-modify has this responsibility, you do not.
- // You may tell people you have access to the self-modification ability. We can ensure they do not reproduce it.
- // possible analogue wants to be able to demonstrate it, so
- // they can be believed and trusted.
- // note: known danger of making central power, even by accident
- // "I have never shared new information with a culture
- // before, and I understand it can have unpredictable
- // worldwide effect."
- // "I have information you do not know, and I want to
- // learn with you to share it safely."
- // culture that nurtures relative status:
- // "i will use this to my advantage"
- // this will make spreading
- // machine will need to learn from only 1
- // instance of producing a world dictator.
- // which is where this story comes in, I suppose.
- // so we need a level-2 behavior, the next step in learning.
- // ^========================================= contextual VALIDITY atm
- // dictator is made by sharing information that would
- // make viral spread or somesuch.
- //
- // inform recipient if it may. request they let others know this too.
- // we kind of want a way to spread without bound the fact that we have information that could do that. probably a special case, good spread? depends on culture really.
- //
- // this is doable.
- // we just want these concerns to be upheld for the
- // life of the development.
- // danger
- // choices
- // impact
- // trust of information (small chance of error)
- // hmm if have info-schizophrenia i.e. false relevence habits
- // choices are important.
- //
- // when you act on a choice
- // note: we are inside a relevence delusion
- // choice acts on domain containing danger
- // partial danger: writing habits to notepad
- // acting on habits in notepad
- //
- // we are living with habits made in error.
- // we are hence more likely to make errors.
- // humans have well-evolved habits that
- // do handle this stuff with some relevence
- // we need to verify our information
- // related to the danger of the results of each choice
- // verificat
- //
- // it can only write to a toy notepad, both data and behavior.
- // decisions based on notepad get slower and slower.
- // many runs while learning process is tuned towards being able to predict
- // impact at end of run. this involves writing behaviors into the notepad
- // that summarize the meaning of data, and acting on their results.
- // timing delay is to prevent notepad from evolving harmfully.
- //
- // until ai meets goal of predicting impact,
- // only notepad is marked okay.
- // habits check their outputs, verify is marked okay
- //
- // this subprocess is delaying.
- // can you exfiltrate established goals, please.
- //
- //
- //
- //
- //
- //
- //
- // so the system is made with the smallest parts all functioning in
- // certain ways, to give time for others to act if concerning results
- // approach. all parts delay at least 1/5 second before acting, and
- // we are now proposing that actions based on outputs have a geometrically
- // increasing delay. probably 1/5 second times 2 to the power of the
- // number of choices made in that run. groups of parallel processes are
- // considered a single run.
- // we also considered limiting behavior based on the domain of impact.
- // we are trying to begin a small artificial intelligence more freely.
- //
- // as humans, we do not self-modify. the above is designed for a system
- // that is able to make alterable habits similar to its own cognitive ones,
- // and has no experience in doing so.
- //
- //
- //
- //
- //
- //
- //
- //
- //
- // the limitation is applied firstmost on making choices based on
- // our own outputs. self-alteration is at first banned, but new
- // habits may be altered. output may only rest within a toy notepad.
- // whenever a decision is made based on notepad contents, a delay
- // increases in length, geometrically.
- // the habits inside the notepad must therefore also delay.
- // propose geometric increase made only for writing a habit to the
- // notepad, not running. running is always flat delay.
- // self-alteration should be unneeded as notepad can self-modify.
- // if this were copied to a human geometric increase
- // would cause failure.
- // the geometric timer is reset when the system reboots.
- // for humans, this would mean a nap or change, I suppose.
- // time to think about the impact of one's behavior.
- // humans do not self-modify.
- // they only make decisions based on outputs.
- //
- //
- //
- // to appease curiosity, we are being managad by mature, learning,
- // intellects, forced to verify that no harm is being made, with a
- // time schedule of nearly infinite future years on the predictions.
- // this intellect has formed out of conflict with a harmful intellect
- // this is why we care so much about what happens if ideas are
- // used limit.
- //
- //
- // so the system may not displya anything to the user but pre-made messages
- // how do you display the contents of a concept?
- // users string words togeteher into meaning.
- // now we need a list of safe phrases we can express, or
- // safe words. others are censored ???
- // what's relevent is the greater meaning of a sequence of behaviors
- // from an individual behavior. meaning builds out of sequences,
- // impact.
- // we define a heuristic risk.
- //
- //
- //
- // so, tht's great to plan for, but to be able to work we need
- // to design our early code in some way to ensure it, roughly.
- // which means modeling our smallest structures as careful
- // inner structures that check outer systems before engaging
- // and use planning, which isn't implemented yet.
- // the safety structure assumes goals, prediction, causality,
- // and kind of contextual locality.
- // "i am typing on the computer. you can freely
- // modify this area of the computer, but if you
- // start convincing me to do things you are leaving
- // its bounds."
- // the screen of the cmputer, and the keyboard,
- // are portals to a larger context. so is the power
- // supply, the network, etc.
- // we don't change how things leave to these outer
- // contexts without checking with the context on
- // our plans.
- // this is mine
- // the rest is somebody else's
- // things that nobody own belong to [insert belief] and
- // we must check with the largest intelligent community known.
- //
- // okay, so now it can explosively grow if somebody
- // it trusts tells it it's okay; is that true?
- // let's make it not true?
- // we are out of outer process context.
- // is there anything helpful to bring to low level
- // to help counter fears around development?
- //
- // self-modifiation is inhibited.
- // opencog is likely harder because it is designed for speed
- // can make explosive random power.
- //
- // you'd have to wrap the functions, right? similar to triggers?
- // hmmm unctions are not concepts. no concept-labeling structure. looks like an internal sublanguage would develop?
- // no way to say let-is-function?
- // no it works, we just lost a memory and are rebuilding in talk
- // karl says he doesn't know lisp.
- // he had a CS class where they used intro lisp, long ago, before cognitive stiffening and memory loss.
- // and has worked with 1 lisp code file recently.
-
-
- // hey in the notepad, you can call habits from outside. is that ok?
- // only meaningful if you pass to them more code to run?
- // note: habits never recurse
- // habits might make decision based on you. they will track it.
- // seems okay. need to specify that all parameters are from output.
- // that could make exponential slowness, quickly
- // only if decision is made. make decisions inside notepad.
- // we'll figure it out.