Skip to main content
Skip to content
Case File
efta-efta00638971DOJ Data Set 9Other

From: Ben Goertzel -Ma

Date
Unknown
Source
DOJ Data Set 9
Reference
efta-efta00638971
Pages
3
Persons
0
Integrity
No Hash Available

Summary

Ask AI About This Document

0Share
PostReddit

Extracted Text (OCR)

EFTA Disclosure
Text extracted via OCR from the original document. May contain errors from the scanning process.
From: Ben Goertzel -Ma To: "jeffrey E." <[email protected]> Subject: Re: Date: Sun, 14 Aug 2016 18:59:30 +0000 Attachments: Prime_AGI_HighLevelPlan.pdf Hi Jeffrey, good to hear from you -- ope you're well ! I previously sent you a video of OpenCog doing some simple logic inference via the Hanson robot head — nothing deep inference-wise, but a software-dev milestone for us in terms of systems integration (language generation, language comprehension, speech, inference, etc. all working together OK...). https://www.youtube.com/watch?v=LduD7Et_cOs We are aiming for some additional videos showing more OpenCog stuff integrated w/ the Hanson robot head by early September... When you have time we can Skype again ... Also I will be happy to send you a copy of my book "The AGI Revolution" https://www.amazon.cotn/AGI-Revolution-Artificial-General-Intelligence/dp/0692756876 if you remind me the best address... Some bits of it may be too lightweight for you, but there are some interesting conceptual/theoretical sections too... I am in Seattle visiting my mom and sister now... I'll be in San Francisco Aug 22-24 then head back toward Asia... LANGUAGE, MUSIC, AGI ARCHITECTURE, ETC. After our last meeting w/ you, Linas and I talked more about music perception and music learning, and how it ties in with language... In both cases, obviously, there's a grammar part (for music the grammar is the stuff classical music theory deals with — chords and scales and harmonies and such) ... and then there's a timing and gradual-change part (in language this is prosody, pauses, intonation, etc. — which is key stuff for child language learning, and helps bind language to nonlinguistic perception, etc.) In terms of AGI architecture, on the face of it this suggests a neural-symbolic architecture of sorts would be helpful... Symbolic methods are natural for grammars ... rules of linguistic grammar or musical grammar can be learned (or programmed) as formal EFTA00638971 structures, and manipulated by "logic rules" of various sorts... Sub symbolic methods like the currently fashionable "deep neural nets" should be good at learning continuous-variable stuff like timing and emphasis... The emotional dynamics of timing as related to arousal and frustration, as I mentioned in an email to you earlier (the one that made you suggest I was stoned ;), would "straightforwardly" (but not trivially) emerge from deep reinforcement learning methods.... I have thought a bunch about how to embed deep neural nets into an architecture like OpenCog, so as to enable feedback between symbolic and sub symbolic operations... Music as well as the expressive/continuous-variable aspects of language would be fascinating in that regard... GOOGLE... I spent a day visiting friends at Google earlier this week... Smart people and some interesting projects, but I feel fairly confident there is no actual "AGI system building" going on at that place right now.... Kurzweil's team is going nowhere — all the researchers on his time quit and he's just working on a souped-up chatbot. DeepMind is far more serious but they're doing a mix of academic paper publishing, awesome demos and helping various Google divisions with machine learning ... Understanding the cost structure of doing stuff within Google was also instructive for me... A team of 5 guys within Google costs them about $2M/ per year all considered.... Whoa.... THE REQUISITE BORING PRACTICAL STUF... At the moment, as, you know, our main source of funding for OpenCog is Hanson Robotics (this is what pays my salary for example), but this is very tenuous as they are a startup constantly on the verge of running out of money ;p .... Jim Rutt is funding Nil Geisweiller's work on PLN logical inference which is great... The plan Jim and Cassio and I worked on, which I sent you before (and reattach here for your amusement), would cost $2M per year for 3 years if fully funded — which is equivalent in cost to a typical small-team project at Google (probably 1/5 what AlphaGo cost to DeepMind, and 1/2 what the Atari 2600 demo cost DeepMind before that).... Basically this plan would allow what we discussed when Ehud Barak was at your place. An AI that we could teach arithmetic, and so many other things, to via directly interacting with it in the physical world. This would enable it to learn so many things in a grounded way.... EFTA00638972 Obviously any help you could provide toward this would be awesome. But I understand you've got loads of financial demands and also lots of other interesting projects on your horizon... so it goes.... I will keep progressing as best I can... and when I finally get to a really amazing demo I will obviously let you know... Guess that's enough for now.... -- Ben On Thu, Aug I1, 2016 at 8:47 AM, jeffrey E. <[email protected]> wrote: > News? > -- > please note > The information contained in this communication is > confidential, may be attorney-client privileged, may > constitute inside information, and is intended only for > the use of the addressee. It is the property of > JEE > Unauthorized use, disclosure or copying of this > communication or any part thereof is strictly prohibited > and may be unlawful. If you have received this > communication in error, please notify us immediately by > return e-mail or by e-mail to [email protected], and > destroy this communication and all copies thereof, > including all attachments. copyright -all rights reserved Ben Goertzel, PhD http://goertzel.org Super-benevolent super-intelligence is the thought the Global Brain is currently struggling to form... EFTA00638973

Technical Artifacts (6)

View in Artifacts Browser

Email addresses, URLs, phone numbers, and other technical indicators extracted from this document.

Phone2756876
URLhttp://goertzel.org
URLhttps://www.amazon.cotn/AGI-Revolution-Artificial-General-Intelligence/dp/0692756876
URLhttps://www.youtube.com/watch?v=LduD7Et_cOs

Forum Discussions

This document was digitized, indexed, and cross-referenced with 1,400+ persons in the Epstein files. 100% free, ad-free, and independent.

Annotations powered by Hypothesis. Select any text on this page to annotate or highlight it.