Skip to main content
Skip to content
Case File
efta-02702508DOJ Data Set 11Other

EFTA02702508

Date
Unknown
Source
DOJ Data Set 11
Reference
efta-02702508
Pages
15
Persons
0
Integrity

Extracted Text (OCR)

EFTA Disclosure
Text extracted via OCR from the original document. May contain errors from the scanning process.
"in the control room of the banquet" Richard P. Gabriel IBM Research Abstract The Turing Test, AI, programming, creativity, and mystery. Categories and Subject Descriptors D.2.10 [Software Engi- neering, Designl; D.3.0 [Programming Languages, General]; 1.2.7 (Artificial Intelligence, Natural Language Processing General Terms Artificial Intelligence, Natural Language Processing Keywords Science; programming; natural language generation I am writing this essay because I am puzzled. In July 2015 I took eighteen haiku-like poems to a writers' conference and presented them as my own work. In reality, a program I created called "Inkwell" wrote them, and I intended to execute a variant of the Thring Test. The results were better than I had hoped for in verifying InkWell as a good poet, but I was left with disquiet about what the experience meant for understanding the Turing Test, programming, the artificial intelligence research program, and what consciousness is. In the Winter of 2014 I programmed my English language revision system [1] to write haiku—just to see whether it could do so plausibly. I let the system run overnight generat- ing about 2000 haiku. Among them were the four at the top of the next column. They stopped me in my tracks because the quick program I wrote was not of the monkeys typing at keyboards variety—instead I programmed the system to determine its own topic and then write coherently about it using a few dozen haiku templates as starting points. And Permisslon to make digital or hard copies of all or part of this work (or petsonal or classroom use is granted without fee provided that copies art not made or distrib• toed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author must be honored. Abstracting with credit is permitted. To con otherwise. or republish. to post on servers or to redistribute to lists, requires prior specific permission andfor a fee. Request permissions from permissionseacm.org, ChswardS14. October 20-2a.201OrPorthodrOvegotRUSA. Consightithekibyths es. h.,.. ..sou.., lh...,.dtu ACM. A04970-1-1503-25tertrlittInAl0. hurifdx.dotorgil0.1145r266ttifir2661I55- deep in the dark— the power of snow walking in the deepness scrupulous in the twilight— the price of gold chases the way of the world in power awake In the dark— the edge of the water can spread in your presence time of life issue: a bird of prey pulls up out of the way into the palm those four haiku are good—not just human-like, but good poetry with two of them close to being exceptional. I worked on the system more over the next six months, broadening and expanding the template language to give more control to InkWell, deepening its understanding of language and the music of language, and adding more observations Ink- Well could make of its drafts and along with them more revisions. Over those months InkWell produced a lot more haiku, and I selected fourteen of them to add to the above four for a Turing Test. In October 1950, the British journal Mind published an essay by Alan M. Turing titled, "Computing Machinery and Intelligence," in which Turing proposed an operational definition for "intelligence" [2]. This definition would come to be called "the Turing Test." Turing himself called it "the imitation game," in which a questioner separated from two contestants would submit questions to those contestants. read their replies, and ultimately choose one as human and the other as machine. Interestingly, Thring introduced this game with a similar but different one in which the interrogator was to attempt to determine the gender of two contestants, one male the other female. Interesting because Turing was homosexual and perhaps accustomed to such an imitation game. But in the matter of intelligence, such an operational definition made some sense. Thring wrote the following: May not machines carry out something which ought to be described as thinking but which is very different from what a man does? This objection is a very strong one, but at least we can say that if nevertheless, a ma- EFTA_R1_02074939 EFTA02702508 chine can be constructed to play the imitation game satisfactorily, we need not be troubled by this objection. -Turing, Computing Machinery and Intelligence, 1950 The Turing Test is at (or near) the heart of the research program called artificial intelligence. In my youth I described artificial intelligence research as an exercise in trying to write programs one doesn't know how to write—at least for engi- neering-type Al research. Some of us generalized this to the idea of "exploratory programming: in which one had a gen- eral sense of what the program should do, and only a partially formulated idea of how to achieve it. In recent years the idea of what programming is has drifted away from including this view toward specifiable. routine. infrastructurish programs and systems. i've heard this referred to as the static-verse. in a famous debate / discussion, Michael Polanyi and Alan Turing talked about whether the mind / the brain was unspecifiable or merely not-yet specified [3]. And what would an incorrect but Thring-Test-passing system be? In his discussion of how the imitation game might go in the computer version, Thring wrote this as the first example of a question in the game: Q: Please write inc a sonnet on the subject of the Forth Bridge. A: Count me out on this one. I never could write poetry. -Turing, Computing Machinery and Intelligence, 1950 In the Summer of 20151 attended the Warren Wilson Alum- ni Writing Conference, which is held annually for graduates of the Warren Wilson MFA program. i am such a graduate. in poetry. The conference was held at Lewis & Clark College in Portland, Oregon. The week was unusually hot and humid for Portland, and this physical difficulty was reflected in an edginess to the conference. My plan was threefold: read the eighteen haiku aloud on the first night of the conference to all attendees: participate a few days later in a writers' workshop as the writer of those eighteen haiku; and on the final day of the conference, give a lecture entitled "is My Program a Better Writer than You?" The abstract for that lecture was as follows: I've been working on a program that thinks like a poet and produces nice stuff. I'll show you how it works and why it's not like the kinds of progmms that do your banking or predict the weather. But everything I'll talk about is really about writing. i read on Sunday night. After the reading a few of the writers came up to me and commented on my reading. My reading was short because the poems were short—and the attendees knew i didn't normally write haiku. Their comments includ- ed these: "terse condensations: "evocative: "took the top of my head or "funny and profound: "natural, personal, and 2 rhythmic: "compact fluid energy: 'wry and elliptical: and "whimsical elegance." I didn't consider this as evidence that InkWell passed the Turing Test. it was the first day of the conference and people were jet-lagged and not entirely prepared for the rigors of the conference; and my reading took about four minutes of an allocated ten. Most writers stretched their reading time at least a little, thus my short reading of short pieces stood out as energetic and sudden. I was still uncertain whether the hint was noticed—the hint contained in the title and abstract for my lecture. Here is what a haiku is: A haiku in English is a very short poem in the Eng- lish language, following to a greater or lesser extent the form and style of the Japanese haiku. A typical haiku is a three-line, quirky observation about a fleeting mo- ment involving nature. -Wikipedia 141 For many, the quintessential haiku poet is Basho in the 17" century; an exemplar of his haiku is the following 151: On a withered branch A crow has alighted: Nightfall in autumn. The nature of haiku is complex and has changed over the centuries—time and place are still essential: counting on is not (some mistakenly conflate on and syllables). InkWell is a small program (about 45,000 lines of Common Lisp code), but it has a lot of data (about 15gb when all the dictionaries, databases, and tables are loaded). Turing wrote "I should be surprised if more than 10' [binary digits] was re- quired for satisfactory playing of the imitation game." InkWell has more than 10". InkWell "knows" a lot about words, per- sonality, sentiment, word noise, rhythm, connotations, and writing. Its vocabulary is probably more than five times larger than yours, gentle reader. The core engine works by taking a template in a domain-specific writing language along with a set of about fifty writing-related parameters and constraints, a description of a writer to imitate, and other hints, and com- piles all that into an optimization problem which the writing engine works to find a good way to express what the template and constraints specify. Although some parts of InkWell were created through machine learning, the overall approach is optimization, not machine-learned transformations. The primary research question is to try to isolate and codify what separates information transfer from beautiful writing. Here is an example of information transfer: EFTA_R1_02074940 EFTA02702509 The summer homes on Long Island were closed. Tonight I watched a ferry begin its crossing to Connecticut. The moon was rising, and as it rose I thought about how the houses are not part of the natural world and what the island looked like to early Dutch sailors coming upon it—like something new. And here is how F. Scott Fitzgerald wrote "the same thing" in `The Great Gatsby" (6J: Most of the big shore places were closed now and there were hardly any lights except the shadowy, moving glow of a ferryboat across the Sound. And as the moon rose higher the inessential houses began to melt away until gradually I became aware of the old island here that flowered once for Dutch sailors' eyes—a fresh, green breast of the new world. -Fitzgerald, The Great Gatsby There is more going on in this version. But what is it? Images, mood, a "vivid and continuous dream" as John Gardner would put it (7j. The first gives us the bits; this one gives us the story. The haiku writer is a driver program that produces the template and constraints the core engine works from. I propose to consider the question, "Can machines think?' -Turing, Computing Machinery and Intelligence, 1950 Turing begins his essay thus. A large question. And largely his essay aims to explore it. The word "think" is the disturb- ing part for many—at least when Turing wrote this essay. Thinking seems like something only humans can do. But even sixty-five years later the full meaning of the word can confuse us. Is thinking puzzle solving, creativity, empathy. wonder, faith, curiosity, ideas, reasoning, reflection, recol- lection, intention, attention, care, imagining, consciousness, language, metaphor, judgment—all of these? Some of these? Turing addresses objections to the idea that machines could think, and offers some suggestions on how to approach achiev- ing mechanical thought. To the objection that only souls can think, Turing asks whether God lacks the power the grant souls to machines. To the objection "I hope not," Baring turns away—though today thinkers like Hawking, Musk, and Gates embrace the fear. To the objection that Godel took care of that for us. TUr- ing points out that G0del took care of only a specific form of incompleteness, which might not be relevant, and besides. why are humans immune from it? To the objection that ma- chines have certain disabilities ("can't feel, can't fall in love. can't make mistakes..."). 'flaring generally derides the idea as not entirely relevant or as not something to be proud of. 3 To the objection that the nervous system is continuous and digital computers discrete. Turing remarks that an interro- gator couldn't take advantage of this because the right sort of answer could be made anyhow by the remote machine. To the objection that humans have informal behavior, 'lur- ing remarks that a machine can easily have laws of behavior, which is really what people have. To the objection of ESP (!!!), Turing admits fear but concludes that a telepathy-proof room will solve that problem. This leave the objections of consciousness and originality. These are subtle aspects of thinking, and though Turning ad- dressed them, I will address them in the context of InkWell, which answers the objections well, and in unexpected ways. The Turing Test is about an interrogator and two subjects: a person and a machine. The test is described as if it happens once, and all the people—and the machine—are ordinary. It doesn't look at extraordinary talents, special skills, and ex- pertise; and the test is presented so that clever avoidance of the question is within the rules. Can the interrogator tell the machine and person apart? Here is your chance to be an interrogator. At the end of this essay in the Appendix is a page called "Thirty-two Haiku." It contains the eighteen haiku I took to the writers' conference. plus fourteen more. Four of the ones InkWell wrote were re- vealed on the first page of this essay, so of the twenty-eight others on that page of thirty-two, half were written by Ink- Well and the other half by Ban'ya Natsuishi, Annie Bachini, and John Ashbery. Have fun deciding which. But the task I just set demonstrates an important problem with the original 'flaring Test viewed sixty-five years after its conception: being unable to distinguish a computer from a person once is not always enough. No one in their right mind and being honest could argue that it's clear which fourteen are which—all the poems seem like they were written by a person or by people. The question is whether there is a dis- tinction to be reasonably observed between those written by a poet and those not. A single non-expert interrogator could easily mistake InklA'ell for a person. Multiple sessions and multiple interrogators are needed. The issue of the proper interrogator has been addressed in the past by pitting an expert human against a computer. If the computer can "defeat" that expert, it has some human- like chops. One of the first examples was the checkers play- ing program written by Arthur Samuel in the late 1950s i8j. It was also one of the first programs to improve itself through machine learning—although a very simple type. The program was able to play advanced amateurs quite well. In 1995, a checkers playing program called "Chinook" won a special checkers championship called "Man-Machine World Championship." (Chinook won the year before, but its creator, Marion Tinsley. withdrew from the competition because of pancreatic cancer.) Chinook had no machine learned aspects. EFTA_R1_02074941 EFTA02702510 In 1997, Garry Kasparov lost to Deep Blue at chess—Kasp- arov was the reigning world chess champion. In one pivotal game Kasparov remarked on the "superior intelligence" of the machine during the first game (won by Kasparov) by avoiding a dangerous position that had short-term advantages; some have reported that this realization shook Kasparov, who lost the second game. And according to other reports, this unusual move turned out to be due to a bug in the software. 'Airing himself created a paper-machine-based chess-playing pro- gram, which Kasparov described as a "competent" player (9j. Beyond chess and checkers are backgammon and other games, which machines are good at. The game Go has re- cently started to succumb to machine play. Away from games, machines have challenged some language-oriented human performances. One of the early examples was PARRY, writ- ten by Kenneth Colby and his students at Stanford Univer- sity (101. PARRY simulated what was then called a paranoid schizophrenic, using a simple model of the condition and a fairly sophisticated English parser. PARRY is considered to be the first program to pass the Turing Test, or a version of it. A group of psychiatrists analyzed a combination of real patients and computers running PARRY through teleprint- ers. Another group of thirty-three psychiatrists were shown transcripts of the conversations. The two groups were then asked to identify which of the "patients" were human and which were computer programs. The psychiatrists were able to make the correct identification only 48% of the time—the same as random guessing. More recent was the IBM Jeopardy!-playing program called "Watson." In early 2011, Watson beat the two of the most suc- cessful contestants on the show, Ken Jennings and Brad Rut- ter. Watson was a stand-alone system with about 3000 cores, 16TB of RAM, and a pretty large store of encyclopedias, dic- tionaries, thesauri, newswire articles, databases, taxonomies, and ontologies—some of which InkWell also uses. Watson was not connected to the Internet. There were many reasons Watson was able to win—some having to do with the Turing Test aspect of the problem, but many having to do with hardware and algorithms. For ex- ample. Watson was routinely able to exploit the difference between humans and the machinery in response speed when the signal was given that "buzzing in" was permitted. Watson was able to use the many previous Jeopardy! games it was tested with to be able to better predict where Daily Doubles were, and it was able to do better betting based on game theory. The software used an ensemble approach that com- bined about a hundred different ways to (statistically) solve the answer, and Watson would buzz in only when there was enough confidence in the early results of this analysis—and it then used the time Alex Trebek used to recognize Watson to continue the analysis. In these experiments, it wasn't an "ordinary" interrogator, an "ordinary" person, and a machine, but expert-level com- petitors, and performance was judged according to difficult 4 standards. Moreover, each of these tests was subject to the Moravec paradox, which states that high-level performance on "intelligent" problems—playing chess or other games, simulating abstract thought, theorem proving, and skill in arenas requiring expertise—is relatively easy to accomplish with not much computational strain, while perception, mo- bility, and other low-level cognitive tasks are comparatively difficult 1111. Moravec and others speculate that long evolu- tionary work developed the latter, while higher cognition appeared recently in animals, and it likely represents a thin veneer on deep sub- and unconscious foundations. The Turing Test is squarely on one side of this paradox. In a writers workshop, a group of writers comment in a loosely structured way on the work brought to the work- shop. The work is distributed well before the workshop so participants can prepare. Each workshop session looks only at the work of one writer. In general, the writer whose work is being discussed remains silent. The comments begin with an overview, then what's good about the work, then how to improve it, and finally the writer can ask questions about the comments. Sometimes there is a teacher or workshop leader; because the Warren Wilson graduates are well-practiced, the workshops at this conference operate without such. My workshop group consisted of four people: CG (woman), MN (woman). DC (man), and me. CG, MN, and DC had five published books of poetry between them, and many magazine publications. I was the last writer workshopped, and my slot was on Wednesday. the day before my lecture. I recorded the workshop. and I will present paraphrases of some of the comments. When I do, I'll use a sans serif font so it's clear it's not a direct quote. In most cases the paraphrases are close to being quotes. DC was the most published of the workshop participants. He started with a "flyover," which is a kind of overview of the work. These are extraordinary and extraordinarily small, large poems. The writer of these—this guy, Richard, or who- ever—he is not a random person, he's not a random guy. I think he understands randomness, so it's all the more scary. He doesn't do things—as a rule—by accident. He makes choices. The variety is amazing on every level: number of syllables, subject matter, syntax, whether they start out specific and go to the general, or start out general and go to the specific. Some of them are simple, some of them are complex, some of them are funny, some of them are dead serious, some are kind-of in the natural world (but mostly not); there are different persons in them; "by myself" is repeated; music seems important. Some are observations, some are moments, some are philosophical and very large (and not just the words, but the ideas). "Murder" is already a big word; EFTA_R1_02074942 EFTA02702511 "murderous" is a bigger word; "murderousness" is about not this fatalist murderousness. as big as you can get. Lots of "ness" deathwatch, words. "Depth" is more boring than but your dead subroutine "deepness: -oc The language of this comment is typical of how working po- ets talk to each other. Notice he said, "The writer of these—this guy, Richard, or whoever...." I asked him about this later and he said that he entertained the idea that my program wrote the haiku, but after considering that for a while, he rejected it as not likely— however, he kept a small hedge. Here's how InkWell produces haiku. The topic for the hai- ku comes from two sources: input from a person and input from one of InkWell's 110 source texts. Input from a person is used if the interactive haiku maker is used; otherwise the topic input comes solely from InkWell's database of texts. I'll describe the two-source process using an example. First a person inputs some words. These words represent a topic suggestion. Suppose the words input are as follows: number, random, player, narrative. InkWell next randomly determines a number of words to select from its textual da- tabase to add to the input words. In this case it decides to choose five words taken from a randomly selected passage from Steinbeck's "The Grapes of Wrath" I12): fire, shifting, rusty, stow, lids. Because they come from a small region in the text, they are not random words—they are related. For each set of words, InkWell constructs a sense, which is a word- vector-like structure from the supplied words and close-by synonyms directed by a complex spreading algorithm which also assigns weights or relevance coefficients to the entries. Then the two senses are combined as follows: cS, S,, where c is a linear factor, S, is the person's input sense, and S, is the sense Inkwell chose from Steinbeck. In this case, c=2.04. The resulting sense (S) can be visualized as a word cloud with the sizes of the words proportional to their associated weights. Appendix Figure I shows the resulting word cloud. The linear factor is always at least 1.0, which has the effect of favoring the person's input. Next. Inkwell chooses a haiku template (the one at the bot- tom of the page, in this case). The template is in a domain- specific language for haiku. This template specifies four senses indicating a season, transformation, completion, and a jour- ney. The sense words are as follows: snow, snowfall, water, ice; fall; complete, finish; and span, bridge. Each of these senses is linearly combined with the sense S above to create the senses that will be used for the haiku. The linear factor for S ranges from 1.0 to 3.0, biased toward 1.0. Combining the same sense with the template senses ensures some degree of coherence throughout the haiku. The next step is to assign random weights to the 32 con- straints InkWell uses for haikus. This includes the language model for InkWell to imitate—in this case it's a collection of my daily poems from 2011 and 2012. InkWell constructs a misfit (unction for these constraints where the function re- turns 0.0 when all the constraints are satisfied. Inkwell se- lects words and phrases to try (28.785 in this case), and then optimizes the misfit function over these choices. A table with all the chosen constraint weights is in Appendex Table 1. In the last step, InkWell reviews the top several haiku for sense (using ngrams) and uses the most best. The final haiku seen just above is not great, but it's an hon- est look at the sorts of haiku InkWell routinely produces. One of the remarkable things about this haiku is that Ink- Well selected the word "underivative" for the specified word "first." This is a choice not many writers or poets would dis- cover. And for technical people the idea of Chalf-randomized number" is interesting. If one were to consider this a poem written by a person, one could analyze it as commenting on how an artificial writer based on random processes could pro- duce a story unlike any seen before. Could a half-randomized number be one produced by an algorithm—a pseudo-random number? I find the more I look at this haiku—which I selected because the parameters it chose illustrated InkWell's writing process even though I didn't like the final haiku—the more meaning and tangents it has. Very human in a spooky way. underivative narrative lighting on the half-randomized number The other two poets made flyover comments; MN remarked: I think he is writing these as a release after a day's work, and they were written over a period of time (not as a group). I see two sorts of language—poetic, concrete language and things in the world, as well as technical or corporate language. It's as if there is a war going on between the two sides of his brain. But the same brain. -MN Here MN reveals she is specifically reading these poems as mine, because she has been in writers' workshops with me before and knows my (real) poetic work as well as my scien- tific work. CG was a little more terse: The language is condensed but plain. ((first adj no-auto-cap) ((local-sense snow) (noun-phenomenon noun-substance) ((snow ice] noun)) (return) ((local-sense falling) verb-weather ((fall] verb) gerund) (return) on the half (word-hyphen) ((local-sense finished) verb-change ((finish complete] verb) past) ((local-sense bridge) noun-artifact ((bridge] noun))) 5 -CG EFTA_R1_02074943 EFTA02702512 this grave— no one sees it mortality, mortality The writers then went on to talk about some of the poems. Images do a lot of work, especially in haiku, and I like to see movement in the haiku, so this one ("this gravel is my favorite—the one I felt so much movement from. This one taught me something, and it changed something. The speaker is in the image even though there is no "I." I even felt the image move. What I learned is that mortality is not just when the body goes, but when the person is no longer remembered. That's just so beautiful. -CG I see it differently. I like all these readings, and I'm a fan of this one too, even if we all read it a little differ- ently. One way is that people don't see the end coming, because they are living their lives and here "mortality" is perking up and saying "don't forget about me"; or also that the writer's current life is like a grave—the daily routine, the getting and spending, and our day-to-day life is a kind of mortality. But this is because of the other poems pointing this way. -DC DC brought up "time of life issue" as one of his favorites. It was from of the original 2000 poems written in 2014. pleasure for me. The talon isn't mentioned, but you can see it; (DC agrees); it isn't mentioned, but you can see it. -CG MN said, "This is a great one? Next, DC brought up "the maiden condominium" as an example of the variety of the poems. It is different from the others. I really like the sound in this one. I don't get the full sense. This doesn't turn me off from being intrigued and trying to understand it. These are big words that have never been put in the same line together before in the history of the English language. (Then he reads the whole poem while CG iaughs.rgametocyte" and "banquet" don't rhyme but they go together. "Ga- metocyte" is a sign of life. (CG and MN repeat "maiden condominium" and "control room of the banquet" and wonder what they could be. They are having fun and laughing.) There is some super power going on in this one. And big words. There is wonderful humor in these. Not standup comic humor, thank God. Not one liners. There is comedy in these. Whimsy. Along with lots of seriousness too. A great combination. Definitely one of my favorites. There is no "I" in it, except there is are "eyes"—someone is observing it, thinking it, and feeling it, and commenting about it. It's time of life Issue: powerful, and it's large and small at the same a bird of prey pulls up time; or general and specific at the same time. out of the way into the palm "time of life issue" could be abstract, but "a bird of prey pulls up" (CG says "wow" in the background) is very vivid and specific, and "out of the way into the palm" is both. It has a sort-of opening up. One of the ways good haiku and short poems work best is they look and feel somewhat tight, concentrated, and highlighted and momentary but there is a kind of opening up—and not just a fly-away, not an escape, necessarily, but an opening up. I feel this; this is a fantastic one. -DC I want to sing the praises of this one too. I want the pleasure of saying how much I like it. Because it took me two or three readings before I got it, before I had an image, and then it was transformer time. You know, everything just transformed. This one shows the power of the form because everything is working together, and I just got a strong image. And it changed, too—it wasn't just given to me. I had to work; there was space in the poem for me. I got the connection and that was the 6 the maiden condominium opens its award-winning gametocyte in the control room of the banquet -DC Then DC quickly mentions "day after day" as another example of good humor. "The maiden condominium" is a good example of something InkWell does well that poets have trouble with. InkWell is relentless in trying to find uncommon things to say and ways of saying things. It's not a coincidence that "gametocyte" and "banquet" almost thyme—Ink- day after day Well uses a concept called echoes to populate poems in the man's can with sonic echoes, a sort of musicality. a man can Poetry seems to be one of the tasks Turing and others con- sider central to the idea of the Turing Test. Recall the first example exchange in a fictional exercise of the test: Q: Please write me a sonnet on the subject of the Forth Bridge. A: Count me out on this one. I never could write poetry. The Forth Bridge is iconic and considered a symbol of Scot- land. In his critique of Thring's idea, Geoffrey Jefferson wrote the following in the British Medical Journal [13]: Not until a machine can write a sonnet or compose a concerto because of thoughts and emotions felt, and EFTA_R1_02074944 EFTA02702513 not by the chance fall of symbols, could we agree that machine equals brain—that is, not only write it but know that it had written it. -Jefferson, The Mind of Mechanical Man, 1949 What does InkWell tell us about this? InkWell selects topics to write about, and then chooses a set of personality traits to display, a set of controlling mood words to use to steer what it says about the topic, and overarching subsenses to direct its inner gaze. Indeed InkWell uses randomness as part of its composition strategy, but as DC pointed out, "<InkWell> is not a random person, <InkWell>'s not a random guy." But does InkWell feel these thoughts and emotions? That's basi- cally what the Thring Test is trying to define. Recent work on consciousness (e.g. "The Ego Tunnel" by Thomas Metzinger [141) has something to say about that, but perhaps the best thought is that in writing this, Jefferson mistakes or misun- derstands the poetic / creative process. Writing a poem is not fundamentally an emotional, expres- sive explosion—it's a deliberate task using practiced skills. It's not Walt Whitman's "1 sound my barbaric yawp over the roofs of the world" [15]. The poem "Howl" by Allen Ginsberg [16] (see Appendix) was mythologized as being a performance piece that was recorded and published (this was part of the testimony at the obscenity trial surrounding the poem), but it was written over a period of nearly two years with critical evaluation by friends brought to bear and specific writing techniques explored and exploited. Ginsberg himself com- mented on the intellectually directed choices and investiga- tions he made while creating the poem. InkWell can be thought of as operating deliberately too. Like Ginsberg, InkWell can decide to experiment with long lines; InkWell can decide the degree and nature of deep in the dark— musicality using rhythms and sounds; InklArell can the power of snow decide to make sense or be crazy; and many other walking in the deepness things like this, but all are deliberate artistic choices. Like real poets, InkWell uses skills to create art. Poets who use feelings alone are the best targets for the criticism of "chance fall of symbols." After Inkwell writes a poem. does it know that it had writ- ten it? In a literal sense it does—it records each poem in a log, sometimes (depending on parameters I set) also noting the artistic choices it made. But in the sense Jefferson meant, no. There is no phenomenal self model in play. That is. Inkwell doesn't maintain an internal representation of what it is do- ing aside from representing its artistic choices. What about the question 'Baring imagines: "Please write me a sonnet on the subject of...." Recall that Inkwell can be directed to look at a topic based on a set of words suggested to it. In Spring 20151 was demoing InkWell to a former long- time colleague; he asked me "can you ask it to write a haiku about this: blues guitar and loud music." I asked InkWell to write five poems, and this was one of them: 7 tuned adrenalin my musk, a beat•boogled heedful I believe this is...well, you decide. CG pointed out one that seemed funny—"the powerful head." DC commented on it as follows: Those words are all deadly—potentially deadly. Unpo- etic, right? They're abstract. Who ever has used "cogni- tion" in a poem? There are some world records being set here. After three lines you realize the poem has turned itself upside down—this poem undercuts itself. Maybe because "powerful head" is already the brain or mind, and it's passing the buck to either itself or some sub-brain or sub-mind, but to support cognition, which means it's thinking about passing the buck on thinking. I didn't want to go there. I'm feeling sorry for whoever is caught up in this (meaning the speaker), because it's just the opposite of what it just said. It's "support cogni- tion," but.. thank goodness I didn't quite go there, even though it wants me to all the time. the powerful head designates its powerful head to support cognition -oc "Deep in the dark" is the first poem to catch my attention from the original 2000 InkWell wrote. The great thing about it ("deep in the dark") I like is that the word "dark" of the first line contrasts with the unexpressed "white" of the snow in the second line. The last line puts them together. -MN I see an echo of "stopping by woods." This is a good echo to have. I really do like "the deepness." It res- cues it. I really can't say why but I know. I tried changing it to "depth." But it's a musical thing or an aural thing. Or "depth" is too familiar and conventional. Each line has a "the" and one could play around with removing them. But removing any of them removes also the par- ticularness of the image. "The" slackens the lines—makes them looser—but it also makes them more immediate, familiar, and more specific -oc According to the most extreme form of this view the only way by which one could be sure that machine thinks is to be the machine and to feel oneself thinking. -Turing, computing Machinery and Intelligence. 1950 EFTA_R1_02074945 EFTA02702514 This is the consciousness argument. In its extreme form the only unequivocal way to look at consciousness is solip- sism—it's just me, babe. But 'Raring rejects that and works toward Jefferson's objection about writing a sonnet by consid- ering whether a viva voce would satisfy him—an oral exam in which the interrogator asks detailed questions about the sonnet.' This leads to this interesting question: to what degree does InkWell know about the poems it writes? Inkwell certainly is not programmed to respond to ques- tions such as "why did you use these particular words right here," but it has an accessible representation of the reasons for all its choices. InkWell decides which artistic choices to make, either through whimsy or by reading a text, how much to weigh them against each other, and which moods or out- side influences to consider. These choices are enshrined in a misfit function InkWell constructs—InkWell composes the source code for this function and then compiles it—and all the choices sit in data structures. You might comment, "Ga- briel, you're exaggerating all this," but these explicit traces are my answer to the difficult question, "how do you debug InkWell?" I need to see how and why all the decisions were made, because the only significant bugs arise from domain- related mistakes, which manifest as surprising utterances and never exceptions or type errors. And to figure them out, I need to examine InkWell's state of mind, as it were. And were I so inclined, I could program InkWell to access more gently this self model when quizzed—more gently than by using data structure inspectors and debuggers. "I chose this pair of words because they sparked off each other well with syllable sounds without being blatant rhymes: because I wanted to come off as extraverted while channeling remorse: because I was trying to include a subtext of explo- ration and discovery. They were also very Hemingwayesque. And the best other choices were these, and they just didn't measure up." Inkwell can't say that, but looking at its param- eters, it's sense descriptions, its halos, its musicality settings. its target personality, the writer's ngrams it's trying to mimic, the recorded results of the component factors measured in InkWell's misfit function, etc. for a particular poem. I can trivially report it. One way to look at it is that InkWell has an effective, op- erational self model, but Inkwell itself is not yet in that self model, and thus Inkwell is only partway toward being con- scious. Inkwell modifies its own self model to change how it makes art. When we "talk" to InkWell about these inner changes and factors, we do so in a nonhuman language. and InkWell responds in the same language. Is this ok? Is this enough? . This is calkd the-Pickwick' lest, becauseTuring's essay describes a series of queslions about Charks Dickcns's "The Pickwick Papers." Sec Appendix. 8 Turing wrote: We also wish to allow the possibility that an engineer or team of engineers may construct a machine which works, but whose manner of operation cannot be sat- isfactorily described by its constructors because they have applied a method which is largely experimental. -Turing, Computing Machinery and Intelligence. 1950 Today we read this as referring to machine learning. Some parts of InkWell are machine learned—judgments about per- sonality and emotions in a text, for example. In some cases these judgments are baked into dictionaries and databases. The central part of InkWell is a meta-heuristic optimization process, the basis of whose operation can be explained, but whose detailed operation in any particular instance is a bit mysterious. The construction of the misfit function is sym- bolic, and unpacking how that function directed the result of the optimization is explainable. That bird of prey poem: I felt a lot of doublenesses, and I love doublenesses. I wouldn't describe it as really dark, even though there is darkness in it. I find it also comical—not really funny. There's whimsy to it, a whimsy tone to it, both. This is a form of doubleness—dark and comical / whimsical—and I don't know how you do d— how you, Richard, do it. This is a very large, small poem. It sounds quiet to me. The last line is not threatening, but the poem starts out threatening. Not to the exclu- sion of others, but this one is really terrific. Raring remarks that Lady Lovelace wrote in her memoir the following (17]: The Analytical Engine has no pretensions to originate anything. It can do whatever we know how to order it to perform. -Countess of Lovelace, Translator's Notes.... 1843 This leads to Thring's "surprise" concern: A better variant of the objection says that a machine can never 'take us by surprise.' -Turing, Computing Machinery and Intelligence. 1950 Everyone—including 'TVring—who has ever programmed has been taken by surprise by their code. And not because of bugs and errors—and not because of randomness particularly. For Inkwell it's because of its relentlessness trying to find in- teresting things, such as the phrase in the King James Bible EFTA_R1_02074946 EFTA02702515 scrupulous in the twilight— the price of gold chases the way of the world in power a reasonable assumption— by myself, sampling in chocolate with the best set of sonic similarities to the noun "computer programmer": "provoked me to anger." I can ask InkWell specifically to surprise me because it has ngrams for millions of texts written since 1890 118) including frequencies of appearance, and I can ask for rare or never-seen combinations. I can ask InkWell to search for unusual syn- onyms; I can ask it to write unlike particular writers. When there are dozens of constraint types with both positive and negative weights, there are few limits to surprise. The poets were surprised, too. CG: "'sampling in chocolate'" is surprising language"; "'guitar-shaped coloring' is surpris- ing. It evokes brown / beige because guitars are made of wood, and it's interesting / surprising that a shape could evoke a color." DC: "I don't know ex- actly the sense of this, but i like the surprise. the sound, the sonic surprise of 'scrupulous in the twilight'"; "i recognize all of it as poetry because of the surprisingness of the language." I asked about the use of language in the poems. CG: "Awe- some." DC: "Good noise. Surprising in a good way. a few days— intriguing. Lots of variety. Big, small, long short, y "Ws° • evaluations with Monte-Carlo rollouts. Our uitar-sha browsing g p loud, soft, complex, plain. Not a single track." ed coloring program AlphaGo integrates these components purpose-built software robots like those that play checkers, chess, backgammon, and Go. These are intended to show that human-level expertise in narrow domains can be exhibited or at least simulated. The Jeopardy!-playing Watson is close to these narrow robots because the domain is trivia—the sort of stuff that Google is good at finding. Watson moves closer to the universal notion of thinking that Turing was approaching. Turing ended his essay with an appeal to a learning ap- proach to get machines close to human abilities. As noted learning is generally taken as machine learning these days. As I write this essay, AlphaGo just marched to victory against a very strong human Go player (Lee Se-dol, a 9-dan profes- sional Go player). As David Silver et al wrote [201: We have developed, for the first time, effective move selection and position evaluation functions for Go, based on deep neural networks that are trained by a novel combination of supervised and reinforcement learning. We have introduced a new search algorithm that successfully combines neural network The Turing Test is of course bogus. At least in the form Tur- ing envisioned. A common strategy for passing is to dodge questions. typically using humor and distractions. Turing's own example shows a dodge as an accept- able response; "Count me out on this one. I never could write poetry." Beginning in 2008 a series of practical Turing tests have been conducted under academic scrutiny, run using the best interpretation of Turing's specifications. In June 2014 an extensive set of interrogations were conducted at the Royal Society [19J. This produced ISO parallel tran- scripts, each of which contains a single interrogator posing questions for five minutes to a human and a chatterbot, with the responses being returned side-by-side at the same time on the same screen. In the Appendix you can see a sample par- allel transcript. In this sample the LHS (left-hand side of the screen) was a female adult human, and the RHS was Eugene, a chatterbot. The judge misjudged the LHS to definitely be a machine and the RHS to be a non-native English speaking human. The judge got it backward. The human on the LHS had weak responses while the machine on the RHS tried to dominate the conversation and was definitely more lively than the LHS. The chatterbot pretending to be Eugene Goostman, a 13-year-old Ukrainian boy, was declared to have passed the Turing Test for having fooled more than 30% of the judges. Part of Thring's idea was that the unexpected scope of ques- tions would be the key to deciding whether the computer was thinking sufficiently like a human. This is in contrast with the 9 together, at scale, in a high-performance tree search engine. -Silver et al, Mastering the Game of Go with Deep Neural Networks and Tree Search, 2016 Here the issue of viva vice comes up—how would AlphaGo explain why it made a particular move? Answers of the form "7 is better than 6" won't work well, but perhaps the people who developed AlphaGo can intuit such answers. AlphaGo lost game four, and here is what was reported in the press 1211: According to tweets from DeepMind founder Denies Hassabis, however, this time AlphaGo really did make mistakes. The Al "thought it was doing well, but got confused on move 8Z" Hassabis said, later clarifying that it made a mistake on move 79 but only realized its error by 87. -httpwwww.theverge.com/2016/3/13/11184328/alphago- deepinind-go-rnatch-4-result And Ttiring would seem to respond when he wrote: May not machines carry out something which ought to be described as thinking but which is very different front what a man does? This objection is a very strong one, but at least we can say that if nevertheless, a ma- chine can be constructed to play the imitation game satisfactorily, we need not be troubled by this objection. -Turing, Computing Machinery and Intelligence, 1950 EFTA_R1_02074947 EFTA02702516 An early goal of artificial intelligence was to understand how people think and act. Most Al research from the 1%0s, 1970s, and 1980s was directed toward symbolic Al, which was writing programs whose inner workings—being directed to- ward emulating thought—could be understood and explained. In the 1980s it became apparent that programs that could do the mental work humans do could form the basis for an in- dustry. Around the same time, progress in machine learning was accelerating alongside advances in computer power, and the idea that it was important to understand how Al programs did what they do was swept aside. One could always fall back on the pop idea that human intellectual performance had an intuitive, only faintly understandable side, along with a deliberate, conscious side—that is, a machine learning side and a symbolic side. This is the heart of the Moravec paradox. Perhaps this pop idea has some merit. After I revealed to the poets in my writers workshop that the poems I presented were actually produced by a program, two of them were good with it though they expressed surprise. The third, though. was quite upset. CG said that it was un- fair for me to keep that information secret—because in such workshops it's assumed implicitly that the work is produced by the writers sitting right there, and all the comments are made with that in mind, including that it is proper to be gentle with those comments. This because all the participants are graduates of the same writing program, and hence are linked by a special. "caring" bond. CG said that perhaps critiques could have been more blunt had the rouse not been in place. I countered by saying that a poem boils down to the words on the page, and everything else is contextual interpretation. We can assume that the person believed to have written the poem is important to its effect on a reader, but it all begins with the words on the page—even if "by the chance fall of symbols' [13j. The upshot of CG's comment is that if it's known that a poem was written by a machine. then the poem's inhuman- ity could be explored and perhaps highlighted. CG said that knowing the haiku were written by a computer would open the door for blunt comments. However, the other side of this coin is more interesting: if it's known that a poem was writ- ten by a human, then the poem's humanity can be explored and highlighted. This is the side of the coin CG said was the default for the workshop. But how does one know that a poem was written by a hu- man? By seeing evidence. The entity claiming authorship looks human, acts human. The Turing Test? Could we say that the Turing Test is what makes us human—at least in the eyes of of other people (or other Thring Test passers)? In this case, aside from my claim of authorship,' the evidence of human 2. Actually, my claim was literally truthful: "Ere been working on short poems recently." 10 authorship are the haiku themselves—forming a feedforward loop in the best case. After some evidence is found that the writer is human, the poem is examined more thoroughly, finding even more evidence of this. Recall what CG said: "... it wasn't just given to me. I had to work; there was space in the poem for me. I got the connection and that was the plea- sure for me. The talon isn't mentioned, but you can see it...." •-••••-• Consider surprise and the Lovelace objection—programs do only what we tell them to do and therefore cannot be considered "Turing human." An extreme form of surprise is for the program to do something far removed from its basic programming. InkWell has never created a recipe nor has it proved a difficult theorem. How could it and why would it? You'd need a program designed to survive and thrive in a dynamic environment to discover and achieve novel capa- bilities. To quote the fictional character from "The Martian" I221, "you solve one problem, and you solve the next one, and then the next. And if you solve enough problems, you get to come home." This is the universal version of the evolution- ary and learning objective function—you reward behaviors and ideas that enable you to live. The environment provides opportunities for learning, and simple, built-in mechanisms ratchet that into new capabilities, habits, and proclivities. This is reminiscent of unsupervised and semi-supervised learning. This is finding patterns in data never seen before; there is no reliable way to do this, but when the learner stumbles across some sort of reinforcement, this can frosted winter, bridge black. turn into a weak form of supervised learning. And in ice white the real world, refinement is always possible. All of this leads to Searle's Chinese Room argument 1231 and the role of consciousness in artificial intelligence. Here is a greatly simplified and distilled rephrasing of Searle's issue: there must be something inside of or causally created by the brain that is operating on meaning and intentions, not sym- bols and syntax; this thing is consciousness. He argues this by putting a human (himself, actually) in place of, essentially, the CPU in an Al program that seems to pass a Turing-like Test in Chinese. He remarks that mechanically processing the rules of Chinese question answering does not constitute "understanding" Chinese because he (Searle) playing the part of the CPU wouldn't: ...it seems to me quite obvious...that I do not under- stand a word of the Chinese stories. I have inputs and outputs that are indistinguishablefrom those of the na- tive Chinese speaker,...but I still understand nothing. For the same reasons, Schank's computer understands nothing of any stories, whether in Chinese, English, or whatever, since in the Chinese case the computer is me, and in cases where the computer is not me, the corn- EFTA_R1_02074948 EFTA02702517 puter has nothing more than I have in the case where 1 understand nothing. -Searle. Minds, Brains, and Programs, 1980 Yet somewhere inside or nearby a person there is a thing or a clutch of things such that if you replace it or those with Searle. Searle would understand what the person does solely by doing what it or they do. I believe Searle is talking about consciousness. But Turing argued that relentlessly seeking out conscious- ness in a program is asking more of the program than we normally ask of other people. As CG might say, we afford people the courtesy of believing they are people. Yet the Turing Test seems to be about finding hints someone is at home in(side) the test subject. The Pickwick test, which Turing made up, is about exploring inner thought processes. Jefferson is after this too. But what of programs? InkWell is the program whose hu- manity we are exploring—does it exhibit traits generally asso- ciated with people, especially creative people who are typically thought of as feeling beings? And if that question even makes sense, is this realm of questions part of a reasonable battery of requirements and specifications? We talk about the "ilities" but are these external and to some extent internal traits part of an "ility" we can discuss? The traits describe being able to act like a human mind, and to be able to answer questions about how and why the program performed certain human mindful behaviors—the Pickwick test. To be able to explain a program's mental behavior probably requires some sort of understandable reification of its innards. And this also implies that the innards of a program are im- portant to its humanity or at least to our understanding of its humanity. This means that this kind of programming is not like typical software engineering, where getting the job done with good performance, correctness, and maintain- ability is paramount. k.a..• •• • In the end a group of expert writers and poets believed that eighteen haiku written by Inkwell were worthy of be- ing considered real and sometimes real good poetry. These writers and poets believed I wrote these haiku, and the ques- tion is how much that mattered. One of three poets in my writers' workshop believed this made a difference, and the other two did not. There also seemed to be opportunities for Turing ratchet- ing: finding some human elements in a poem increases the 11 feeling that the poet is a real person, and that increases the likelihood of finding more and deeper human elements in the poem—or at least increases the incentive to look for hints of humanity and the energy to look with. In writing InkWell I was trying to explore how poets and "real" writers write—I was trying to capture in a program what I had learned while studying how to write poetry. I in- tentionally started at the word-choice end of things where many of the effects writers use are hidden in sound (noise). connotation, mood, author personality, and influence. By pursuing this I intentionally made the internals of Inkwell as expressive as was reasonable so I could see the effects of changes and additions, as well as study how the different in- tended effects interacted to produce different texts. Neverthe- less, InkWell has many learned aspects—machine learned at my hands, machine learned by others, and learned through curated and automatically produced dictionaries and data- bases. The interplay between the symbolic and learned aspects of InkWell can be observed decently well, and perhaps the nature of this interplay could provide useful research results into the possible nature of the mind. I have written a poem every day since March 18, 2000. That's a lot of poems. Some nights when I sit down to write my daily poem, I "don't have it," as they say. My talent has taken the night off, nothing happened during the day to serve as a trigger, or I'm simply a little too fatigued to crush it. For the past few years when this happens, I have turned to InkWell to help me. I tell it some odd topics to consider and ask it to write a few dozen poems. And from those I'll revise to a good poem or will use the sequence as a starting place. InkWell is a good helper. The Thring Test provides an interesting lens or instrument useful for exploring what we can make of the semi-living na- ture of programs that are designed for a little bit more than their useful effects. We could go crazy exploring all the ins and outs of philosophy of mind, strong and weak AI, con- sciousness, machine learning versus symbolic deliberation, and intuition versus reasoning. but all I'm wondering about is the lesson to learn from a group of hardcore poets taking Inkwell as a colleague. EFTA_R1_02074949 EFTA02702518 References IIJ Richard P. Gabriel, Jilin Chen, Jeffrey Nichols, InkWell: A Creative Writer's Creative Assistant, Creativity & Cog- nition, 2015. Pi Alan M. Turing, Computing Machinery and Intelligence, Mind, October 1950. (3] Alan M. Thring et al, The Mind and the Computing Ma- chine, The Rutherford Journal. 1947. httpt://en.wikipedia.org/wiki/Haiku Robert Han, editor, "The Essential Haiku: Versions of Bubo, Buson, & Issa," The Ecco Press, 1995. F. Scott Fitzgerald, "The Great Canby," Charles Scrib- ner's Sons, 1925. John Gardner, "The Art of Fiction: Notes on Craft for Young Writers," Vintage. 1991. Arthur Samuel. Some studies in machine learning using the game of checkers, IBM Journal of Research and De- velopment Volume 44, Issue: 1.2. 2000. Garry Kasparov, The Chess Master and the Computer, review of Chess Metaphors: Artificial intelligence and the Human Mind by Diego Rasskin-Gutman, in New York Review of Books, Volume 57, Number 2, Febru- ary 11, 2010. (101 Kenneth Colby, Turing-like Indistinguishability Tests for the Validation of a Computer Simulation of Paranoid Pro- cesses, Artif. Intel]. 3(1-3) 1972. pp. 199-221. HI] Hans Moravec, Mind Children, Harvard University Press, 1988. (12] John Steinbeck, "The Grapes of Wrath," The Viking Press, 1939. (5] (61 (71 (81 (9] 1131 Geoffrey Jefferson, The Mind of Mechanical Man. British Medical Journal, 1949. 1141 Thomas Metzinger, "The Ego Tunnel: The Science of the Mind and the Myth of the Self," Basic Books, 2010. DM Walt Whitman, Song of Myself. "Leaves of Grass," self- published, 1855. (161 Allen Ginsberg, "Howl and Other Poems," City Lights. 1956. (171 Luigi Federico Menabrea, Ada Lovelace, Sketch of the Analytical Engine invented by Charles Babbage...with notes by the translator, translated by Ada Lovelace, in Richard Taylor, "Scientific Memoirs 3," Richard and John E. Taylor. pp. 666-731, 1843. (181 http://storage.googleapis.comThooksMgrams/books/datasetsv2. html 1191 Kevin Warwick & Huma Shah. Can machines think? A report on Turing test experiments at the Royal Society, Journal of Experimental & Theoretical Artificial Intel- ligence. 2015. (20] David Silver et al. Mastering the Game of Go with Deep Neural Networks and Tree Search, Nature 529, pp.484- 489, 28 January 2016. (211 http://www.theverge.com/2016/3/13/11184328/alphago-deep- mind-go-match-4-result (221 The Martian (2015 movie). http://www.imdb.comititlei U36593881 (231 John Searle. Minds, Brains, and Programs, Behavioral and Brain Sciences. 1980. EFTA_R1_02074950 EFTA02702519 Appendix deep in the dark— the power of snow walking in the deepness I'm not the same on an island with destructive rhythm behind a rock on the green slope dead soldiers' spirits the powerful head designates its powerful head to support cognition this grave— no one sees it mortality, mortality shopping parade— people step over the broken cassette a bitch. this deep in trick a fortiori not a man not this fatalist murderousness, deathwatch. but your dead subroutine time of life issue: a bird of prey pulls up out of the way Into the palm awake in the dark— the edge of the water can spread in your presence day after day in the man's can a man can scrupulous in the twilight— the price of gold chases the way of the world in power tuned adrenalin my music. a beat-boogied headful under the sea a fish becomes human In an air pocket a crooked rag day— by myself dunking distracted sardines bare branches. tonight again stars, stars arc misprints Thirty-two Haiku an on-the-far-side summer night— whipping up high tea. [1276911 we stripped pickles 14242311 aboard a boat a round table dancing— (467771) an old song [2665271 shoved off the stairs— falling I become [4833111 a rainbow [359659) the maiden condominium opens its award-winning gametocyte [3660191 in the control room of the banquet 1238801) [4833371 [3571911 [263573) [357781) a reasonable assumption— by myself, sampling in chocolate from the boulder smiling up at heaven the continent begins old lift: through the grille three women in pastel t-shirts rural signal, cannot understand Oregon —agricultural 13635891 1159947) [1733171 13844731 parted in the middle— the authority of the air conditioner (1087911 perfection In the brightness 11356971 too late: the last express passes through [306473) the dust of gardens 14907571 [471853) [3030191 [275309) (2656511 a blue anchor— grains of grit in a tall sky sewing pirates Imitate the ways of ordinary people myself for instance dental hospital— dead files line the light casings the hostile defense leads its problematic rear, the rear of frustration a few days— I361597) [102941] 12592831 13482091 by myself, [494497) browsing guitar-shaped coloring 1160967) (269239) in rags and crystals. sometimes with a shred of sense an odd dignity Haiku labeled with primes were written by Inkwell 13 13498231 EFTA_R1_02074951 EFTA02702520 .. ... 7 m rc cornetist Tri ... 01 ° atomic-number" m fibonacci-number . i,_ 2_.. I temente scorer i „ d folktale . r preyalenceadsup 4 b_ numerousness ”., complex-quanrinc1/10iinist or-miss u a ,_unselected ognksman w 4- clarinetist lutanist t at, n: C. .... §:violoncellist> .- ›..'6..L. trumpeter. _ 0 ,,, shooter 1 ..o.c >r e a stringer .- e ac o palpebraumplano-player participanO (=Go) prinitcluantity communicatory t !LI 0, 13).—alcard—player sorft-"" e rhyg w = rix O- isoccer- playerwolistom ic-t hockey-player floating-point-numberree. 2 u C fiddler ' ° V,hornist SOD-SION - 47, c-6 billiard-player oxidation-numbet 0, .w iso _ v i vjaavolleyball-playerlutist fl c— r° gam ist to E. recorder-player ,..2 c i_.0 luirawl pan om Clarinettist keyboardist 2most-valuable-player2 E 3-5, jazzim cea-nhol, e 2 football-player. B y- rilayerMUSiCiall shifting r-.it t baseball-player c worii-our accompanist tlergodic nurn rharpist rhythm-and-blues-musicianoymerositym 2 .- 1- payer -player constant ta: d lacrosse-player rock-'n'-roll-musician 2 bowl-- 41 4- t. E ci":3- 71:11 g-xnarra 'vets. rat ig a er e die?' ict esri Ortis t I °>-...- ° bell-ringerws chess-player o. count1,1 anClt iti cSieaend C 0 u ballplayerin k to player E o - grandmaster- sitar-player porous-wove ru — .t ibistguitar -player .E bassooniis$ it ire 6:. harm . oniser= „,. 1,_ oo / Mal er binist 'wore osavxopjf . `0::: -_ :electric-2Ow w IfittJadrate serverwtcod u r ionist._= ,y,6 onist w aim cro to .0 a a , e r i 0 O. 0, — IV -0 ecimal °pianist er an sman 0, or „s o i. cirri t m golfer i=2 1tn nu r muitiplier-facror prim mu =risKer w v, . = re piper soloist fourth-Power.° majority — Appendix al mvp oboist — ° Appendix Figure 1 Agreeableness 10.0 Conscientiousness -10 Extraversion 0.0 Neuroticism 10.0 Openness 10.0 Openness-To-Change 0.0 Big5 Strength 3.19 Common Word Bonus 2.90 Global Halo Frost.halo Global Halo Bonus 434 Local Halo Bonus 5.76 Local Sense Bonus 40.00 Local Predicates Bonus 938 Synonym Proximity Bonus 4.97 Constraint Bonus 3.41 Rhyme Bonus 47.21 Howl, Allen Ginsberg, first three lines Echo Bonus 37.29 Synonym Search Depth 2 <choose> Search Depth 1 Synonym Decay Rate 0.57 Maximum Synonym Sense,. 2 2-gram-bonus 75 3-gram-bonus 85 4-gram-bonus 85 5-gram-bonus 75 Writer-2-gram-bonus 27.698803 Writer-3-gram-bonus 45.65159 Writer-4-gram-bonus 71.91725 Writer-5-gram-bonus 44.505795 Writer File Poemsrpg.txt Long Words True Appendix Table I I saw the best minds of my generation destroyed by madness, starving hysterical naked. dragging themselves through the negro streets at dawn looking for an angry fix. angelheaded hipsters burning for the ancient heavenly connection to the starry dynamo in the machinery of night.... 14 EFTA_R1_02074952 EFTA02702521 Royal Society Turing Test, June 2014 Transcript I -LHS (16:20:48) Judge what isyour worst memory in your life so far? (16:21:06) Entity: I don't understand. Explain. (16:21:35) Judge: memory of something that happened to you that was bad or painful (16:22:581 Entity: Hmm that a tricky question. My life is won- derful. I overcome many bad things that have happened to me in the past. However, the most painful memory I have had is performing very poorly in my GCSE's when I was younger. (16:23:561 Judge: did you fail? (16:24:12) Entity: Partially. (16:24:35) Judge: explain? (16:25:04) Entity: Some grades turned out to be good. Some turned out to be quite bad. (16:25:37) Judge: and that is the worst thing that has happened to you so far? Mmm how old are you? The Pickwick Test Transcript I—RHS (16:21:00) Judge: what did you have for breakfast today (16:21:10) Entity: Is it 'the third degree'? I'm not going to answer such a provocative question! Well, let's go on though. (16:21:56( fudge: do go on...your call? (16:22:06) Entity: It is highly possible. By the way, what's your occupation? I mean—could you tell me about your work? (16:22:30( Judge: so you are asking me now eh? I am an actress and writer (16:22:381 Entity: Well, I'll read something of your books. if they ain't too tedious (which is highly possible:-) (16:23:151 ►udge: grammer...aren't too tedious! A bit rude of you... (16:23:17) Entity: Calm down, please. (16:23:49) Judge: i asure you i am very calm ... are you? (16:24:01( Entity: Don't even bother me with the fact that you are very calm ... are mine. And I forgot to ask you where you are from ... (16:25:01) Judge: I am from Dorset...where are you from? (16:25:08) Entity: I am from Ukraine, from the city called Odessa. You might have heard about it. Interrogator. In the first line of your sonnet which reads "Shall I compare thee to a summer's day," would not "a spring day" do as well or better? Witness: It wouldn't scan. Interrogator: How about "a winter's day." That would scan all right. Witness: Yes, but nobody wants to be compared to a winter's day. Interrogator: Would you say Mr. Pickwick reminded you of Christmas? Witness: In a way. Interrogator. Yet Christmas is a winter's day. and I do not think Mr. Pickwick would mind the comparison. Witness: I don't think you're serious. By a winter's day one means a typical winter's day, rather than a special one like Christmas. t5 EFTA_R1_02074953 EFTA02702522

Related Documents (6)

Dept. of JusticeAug 22, 2017

4 MAY 11 1241-MAY 12 219_Redacted.pdf

Kristen M. Simkins From: Sent: To: Subject: Attachments: C. Kay Woodring Wednesday, May 11, 2016 12:41 PM Bryan L. Sampsel; Denise L. Elbell; Eileen B. Mckinney: Jeffrey T. Hite: Jonathan D. Grine, Judge:Joseph S. Koleno: Mark Higgins; Melanie L. Gordon: Michael Pipe: Richard C. Smith; Stacy Parks Miller, Steve Dershem daily pop reportdocx daily pop reportdocx CORRECTIONAL 700 Rishel Hill Road Richard C. Smith, AIS CCHF Bellefonte. 16823 Telephone (814) 355-6794 Fax (314) 543.1150

98p
Court UnsealedDepositionFeb 5, 2026

P009747-020326-001-335

March 19, 2025 Mr. Russell Brown Executive Director Washington Association of Prosecuting Attorneys 206 10th Ave SE Olympia WA 98501 Subject: Update - Potential Disclosures concerning Government Witness – CVEO 1 Leonardo J. Bobadilla, Badge Number X807 Dear Mr. Brown: On April 17, 2024, the Washington State Patrol (WSP) provided notification regarding the above￾named employee: “It is alleged CVEO 1 coerced and/or assaulted his ex-girlfriend. It is also alleged CVEO 1 engaged in acts of child

335p
Dept. of JusticeAug 22, 2017

1 May 1 1255-May 6 237_Redacted.pdf

Kristen M. Simkins me: Sent Tn: Subject: Atladimem: LT. THOMAS E. ALLEN JR Thomas S. Allen. Jr. Sunday. May BIL EDIE 12:55 AM Allyson FL Dwell; Brenda McKin1e?c C. Kay Wandring: Caitlyn D. Neff: Daniel?le Minarch?lck: JeFFrey' T. Hite; Jon D. Fisher. Jonathan M. Mfl?n-der. Joseph 5. Kolenorluan Mendez: Kevin T. Jeirles; [any Lidgett Lee R. Shea??er: Lorinda L. Brown.- Matti-new T. Fishet: Melanie Gordan; Michael S. Woods Richard C. 5mm; Shephanie D. Calander?mtus Report SMDIE 20150501004

493p
DOJ Data Set 10OtherUnknown

EFTA02064020

3p
Dept. of JusticeApr 9, 2021

Epstein Palm Beach Property Sale

IN THE SUPERIOR COURT OF THE VIRGIN ISLANDS DIVISION OF ST. THOMAS AND ST. JOHN ******************************** IN THE MATTER OF THE ESTATE OF ) JEFFREY E. EPSTEIN, ) PROBATE NO. ST-19-PB-80 ) Deceased. ) ACTION FOR TESTATE __________________________________________) ADMINISTRATION NOTICE OF SALE COME NOW the Co-Executors of the Estate of Jeffrey E. Epstein (the “Estate”), DARREN K. INDYKE and RICHARD D. KAHN, and hereby file with this Honorable Court this Notice of Sale in the above captione

33p
Court UnsealedMay 7, 2017

PP B0078 Nitro W. Va. plaintiffs' response to Monsanto motion for summary judgment

IN THE UNITED STATES DISTRICT COURT FOR THE SOUTHERN DISTRICT OF WEST VIRGINIA BROWN, ET AL, plaintiffs, VS. MONSANTO COMPANY, Delaware corporation, Defendant. CHESTER A. JEFFERS, Plaintiff, vs. MONSANTO COMPANY, a Delaware corporation, Defendant. AT CHARLESTON CIVIL ACTION NO. 81-2239 DOCKET NO. 225 RESPONSE TO MOTION OF DEFENDANT, MONSANTO COMPANY, FOR SUMMARY JUDGMENT BASED ON STATUTES OF LIMITATIONS Paul L. Pratt, Esquire Route 111 At Airline Drive East Alton, IL 62024 Telephone:

109p

Forum Discussions

This document was digitized, indexed, and cross-referenced with 1,400+ persons in the Epstein files. 100% free, ad-free, and independent.

Annotations powered by Hypothesis. Select any text on this page to annotate or highlight it.