Robot inventors, Robot Patents, Robot Examiners, and Robot Patent Prosecutors
By
By
Henry H. Perritt, Jr.[1]*
The United States patent system is a strong candidate for wider deployment of generative AI. Even though the Patent Office so far has rejected the idea, generative AI systems are on the verge of being able to invent and thus qualify to be listed as inventors on patents. Specialized generative AI engines themselves qualify as patentable inventions, and the Patent Office has granted more than two dozen patents on generative AI. It has pending more than 100,000 applications that use the term “artificial intelligence.” Apart from the patentability of AI technology and its work product, generative AI already is a mainstay of patent prosecution, performing prior-art searches, drafting patent applications, checking their formats, and producing drawings. The power of the technology makes each of these changes in the way the patent system operates inevitable, but each presents legal and managerial issues. In particular, wider use of AI may cause the number of applications to proliferate, cause more of them to be submitted and prosecuted without the aid of skilled patent lawyers and marginalize human involvement at both the practitioner and examiner level. Early in 2024, USPTO issued policy guidance on AI inventorship and on use of AI in patent prosecution. Later in 2024 USPTO invited comment on the impact of generative AI on prior art. Controversy continues over the desirability of patents at all in the present legal regime. Justifications for defense of the status quo, particularly for the rule that only natural persons can be inventors, is thin, unpersuasive, and inconsistent with the way copyright law treats individuals when their creative work is sponsored by inanimate business entities.
Lee Dresden and Kayden Miller were playing their weekly and very competitive game of squash at the YMCA’s Arlington, Virginia, Tennis and Racquet Club, not far from River House, at 1111 Army Navy Drive, where the two shared an apartment. Kayden was leading 10 to 9. Kayden had hit a dropping shot off the right front corner and Lee scrambled to reach it. He managed a looping shot off the middle of the front wall. He was sure that Kayden would return it to the opposite corner and would win the game. Instead, Kayden hit it hard, smacking Lee squarely in his right buttock. The hard little rubber ball stung like crazy. Whirling around, Lee demanded, “Why did you do that?”
Kayden was doubled over in laughter. “I just couldn’t resist. I decided that was better than winning the game. It was.”
“Hitting you with my racquet would be even more fun.”
“Calm down. Let’s go have a beer.”
Lee and Kayden had been classmates and Sigma Alpha Epsilon (“SAE”) fraternity brothers at MIT, both majoring in Course 6-2, electrical engineering and computer science. After they graduated, Lee went to work for the National Security Agency in Fort Meade, Maryland, and Kayden started law school at George Washington University in the part-time program while he worked for the United States Patent and Trademark Office (“USPTO”) as a patent examiner. He chose that arrangement because neither he nor his family had substantial financial resources, and he did not want to burden himself with the hundreds of thousands of dollars in loans a law school like Harvard would have saddled him with.
Kayden recently graduated, passed both the Virginia bar and the patent bar, and went to work for the prestigious Georgetown law firm specializing in patent law, Milligan, Anderson, Jarrabow, and Ferret. Lee, restless under the endless self-protective bureaucracy of the NSA, resigned at the three-year mark and joined a small startup, Vater LLC, headquartered in Manassas, Virginia. Vater had raised substantial capital to develop generative AI solutions for industry.
“What’re you working on?” Kayden asked after they had showered, changed clothes, and had beers in front of them. He laughed. “For the longest time, I couldn’t ask you that—or there was no point asking, because you wouldn’t answer, but I trust that Vater LLC is not secretly intercepting Vladimir Putin’s cellphone calls. Anyway, you can trust me because I persuaded the firm to accept you as a client—even though no one there, including me, thinks you’re ever going to make anything of yourself.”
“Ha!” Lee said. “Trust you after what you just did to me on the squash court? I should know better, but I guess I do need your services, for what they’re worth.”
“I’ll have you know they’re billing me out at $800 an hour,” Kayden said.
“What you rapacious lawyers charge has nothing to do with what you’re worth.”
“Stop the foolishness and tell me what you want my help on.”
“Our new generative AI system, which we have named, ‘Edison,’ has invented an agricultural cherry picker. We want to apply for a patent in its name.”
“What? The Patent Office doesn’t give patents to machines.”
“It should in this case,” Lee said. “Edison did all the work. Let me show you.” He pulled a couple of sheets of paper from his backpack and unfolded them on the table.[2] “As you can see, all I did was ask Edison to identify fruitful areas for invention, pick one of the areas it suggested, then asked it to invent something, and finally asked it to draft a patent application for its invention.”
Kayden looked over the transcript. “That’s not much of a patent application,” he said. “It’s pretty generic–it’s completely generic. Patent applications have to be specific and definite.”
“You didn’t read it all,” Lee said. “Flip the page. I asked it to fill in details.”
Kayden flipped the page and stared at Lee.
“Trying to calculate how much $800 an hour amounts to for only five or ten minutes? Tune it up and submit it.”
“I don’t know,” Kayden said. “I would be more comfortable if you and the other nuts at Vater wanted to get a patent on Edison instead of on Edison’s work as an inventor.”
“I suppose that’s a possibility as well,” Lee said. “The venture capitalists are drenching us in money, and we have to spend it on something.”
“It’s not like this is a complete surprise,” Kayden said. “Most of the hype about generative AI is nonsense, but it sure is changing the practice of patent law. The law firm regularly uses two commercial AI products to lighten the load of drafting patent applications and to search for prior art that might show an invention is not novel and therefore not qualified for a patent. On the last couple of applications I’ve done, I didn’t have to do much more than dictate a fairly high level description, and the software drafted the claims, wrote the rest of the specification, the abstract, and even did a first cut of the drawings. The documents obviously weren’t ready for submission to the Patent Office, but it only took me a couple of hours to refine them and then send them off. The way lawyers send things to the Patent Office these days is all electronic—the Patent Office’s Web-based Patent Center checks the submission for completeness and correct formats and kicks it back with detailed, paragraph-by-paragraph feedback if the submission has errors.”[3]
“The patent examiners long have had a huge stock of standard form letters to communicate various kinds of objections and rejections to patent claims. Based on my experience with some applications and examiners, I suspect there’s no human examiner there, just a bot claiming to receive an application, to examine it, and then to reject it by sending a standard form letter.”
Lee laughed. “Does it pass the Turing Test?”[4]
“What? Oh you mean –”
“Right, you communicate with something that is shrouded so you cannot see it and assess whether it’s a human being or a machine by the nature of its responses to your questions. The machine is intelligent when you can’t tell the difference between its responses and those of a human being.”
Kayden snickered and then started laughing so hard he knocked over his beer. Quickly grabbing it, he gasped, “Suppose we get a patent in Edison’s name, who would own it?”
“Edison, obviously.”
“No. Edison, however smart he pretends to be, is not a legal person. Vater LLC might own it because it owns Edison.”
“Or OpenAI or Sam Altman,” Lee said. “Edison is built on top of ChatGPT, which is a product of OpenAI.”[5]
“I don’t think so,” Kayden said. “OpenAI’s terms of service say that input to and output from ChatGPT both belong to the user.”[6]
Lee looked at him. “So this idea of something like Edison being an inventor is not the first time you’ve thought about this.”
As fictional Lee Dresden and Kayden Miller observe, generative AI can be especially productive when it is applied to patent law. In the business jargon of the day, the “use case” for patent AI is strong. Approximately 350,000 patents are granted in the United States each year.[7] USPTO received 457,500 new patent applications in FY 2022, an increase of 1.6% over the number received in FY 2021. The inventory of unexamined patent applications is anticipated to be approximately 694,600 for FY 2023, an increase of approximately 5,100 from FY 2022.[8] USPTO has a patent examiner corps numbering roughly 8,500.[9] Any activity with that kind of volume begs for computerization and, when it involves highly structured information formats, as the patent system does, it makes computerization easier.
The United States Patent and Trademark Office has been a pioneer in government automation. Its database of issued patents and patent applications was one of the earliest large-scale systems to manage government information in electronic form, spawning some controversies over public access in which this author was a participant.[10] Now, USPTO has implemented its Patent Center, which requires patent applications and supporting documentation to be submitted electronically,[11] and has deployed a number of tools automating the functions of patent examiners.[12] Patent practitioners routinely use sophisticated computer software applications that assist in searching for prior art that can defeat the novelty required for a patent, and that automate the routine parts of application drafting and responding to office actions by examiners.[13]
The excitement about generative AI and highly publicized efforts by at least one applicant to apply for a patent naming a computer program as the inventor has focused attention on the relationship between AI and patent law.[14] More broadly, a few industry leaders and activists claim that generative AI is an existential threat to human society. In fact, the world is not about to come to an end, and civilization as we know it is not going to be brought to its knees by this latest increment in natural language processing.
Nevertheless, despite the unlikelihood of societal revolution, generative AI has the potential to revise significantly the way patent law is practiced, patent applications are reviewed, and patents are issued and enforced.
Part II, following this Introduction, provides an overview of patent law and then explains generative AI.
Part III explores the role of generative AI systems as inventors, as subjects of patents, as aids in patent prosecution, as aides to patent examiners, and as stimuli for more pro-se applications. Many of the results reported in this part arose from the author’s own use of AI-enabled tools.
Part IV asks whether the changes likely to result from the uptake of generative AI matter.
An Appendix reproduces the transcript of an invitation for a generative AI system to create an invention and to draft a patent application for it.
The patent system is international in scope. This article focuses on U.S. patent law only because extending beyond American borders would needlessly complicate it and because the U.S. is a reasonable model for the rest of the world regarding the relationship between AI and patents.
Patent law is a bargain struck between opponents of crown monopolies that proliferated under Stuart kings and proponents of industrial progress who wanted incentives to encourage inventors and to induce them to make their secret inventions available to everyone.[15] The result was the Patents and Copyrights Clause, Article 1, Section 8 of the United States Constitution, which invites the Congress “[t]o promote the Progress of Science and the useful Arts, by securing for limited Times to . . . Inventors the exclusive Right to their respective . . . Discoveries.”[16]
The Congress accepted that invitation in its very first session by enacting the Patent Act of 1790,[17] which authorized a committee of the cabinet, the Secretary of State, the Secretary of War, and the Attorney General, to grant patents “if they shall deem the invention or discovery sufficiently useful and important.”[18]
The statute was replaced in 1793,[19] to define eligible subject matter and to vest authority in the Secretary of State alone;[20] and amended in 1836,[21] to establish an examination system by a federal agency called the United States Patent Office;[22] in 1870,[23] to codify a number of separate statutes, among other things, transferring the Patent Office to the Department of the Interior,[24] to require that examiners be technically skilled,[25] requiring applicants to submit written descriptions,[26] and providing for printing of patents and those descriptions for use of the public.[27]
The Patent Act of 1952,[28] codified the requirement that subject matter be “inventive” in order to qualify for a patent, now known as the nonobvious requirement.[29] Most recently, the America Invents Act of 2011 repealed the traditional first-to-invent system of priority with a new first-inventor-to-file system similar to that in use by the rest of the world.[30]
The substantive requirements have not changed much over the life of the different patent statutes. All of the statutory requirements are aimed at protecting the public domain against unnecessary enclosure by assuring that patents are granted only for certain subject matter, and for inventions that are novel and nonobvious. Inventors must set forth the metes and bounds of the property interest they seek clearly and do so in a way that enables others to practice their invention once their monopoly expires.
Under current law, the subject matter of an invention must qualify for a patent under section 101.[31] Even when an invention involves a process, machine, manufacture, or composition of matter—the categories enumerated in section 101—it nevertheless is non-patentable if it falls within one of the judicial exceptions to section 101: if it comprises a law of nature, a natural phenomenon, or an abstract idea.[32] In Alice Corporation Party, Ltd. v. CLS Bank International,[33] the Supreme Court reiterated these three “judicial exceptions,” noting that they have applied for “more than 150 years.”[34]
Applying the judicial exceptions is known as Step I[35] of the Alice[36]/Mayo[37] analysis. Even if an invention flunks Step I of Alice/Mayo, it nevertheless may be patentable if it includes something else, an extra step beyond the mere judicial exception.[38] This is Step II of the Alice/Mayo analysis.[39]
The concern that animates the judicial exceptions is preemption: upholding a patent covering these exceptions,
would pre-empt use of this approach in all fields, and would effectively grant a monopoly over an abstract idea . . . . Laws of nature, natural phenomena, and abstract ideas are the basic tools of scientific and technological work. Monopolization of those tools through the grant of a patent might tend to impede innovation more than it would tend to promote it, thereby thwarting the primary object of the patent laws.[40]
After quoting Mayo on the preemption concern, the Court then described and embraced the two-step analysis articulated two years before in Mayo:
First, we determine whether the claims at issue are directed to one of those patent-ineligible concepts. If so, we then ask, “[w]hat else is there in the claims before us?” To answer that question, we consider the elements of each claim both individually and “as an ordered combination” to determine whether the additional elements “transform the nature of the claim” into a patent-eligible application. We have described step two of this analysis as a search for an “inventive concept”—i.e., an element or combination of elements that is “sufficient to ensure that the patent in practice amounts to significantly more than a patent upon the ineligible concept itself.”[41]
Patent applications typically attempt to come within the ambit of the text of section 101 by using one of the magic words: process, machine, manufacture, or composition of matter.[42] They typically try to bypass Step I of Alice/Mayo or satisfy Step II by identifying concrete physical or conceptual steps in a method or concrete features of an apparatus that distinguish it from an abstract algorithm or law of nature.[43]
The Patent Office published guidance on subject matter eligibility under section 101 in January 2019.[44] The Office published examples to be used with the 2019 guidance.[45] Example 39 pertains to an AI invention.[46]
To qualify for a patent, inventions also must be novel under section 102.[47] This means that they must not be anticipated by any prior art reference.[48] A prior art reference defeating novelty is said to anticipate the invention.[49] The statute defines prior art as “patented, described in a printed publication, or in public use, on sale, or otherwise available to the public.”[50] In Helsinn,[51] the Supreme Court made it clear that secret sales may nonetheless constitute prior art.[52]
Novelty is defeated when a single prior art reference contains every element of the claimed invention.[53] An initial screening for such anticipatory prior art can be and frequently is conducted by constructing natural language search terms made up of language fragments from the claims in the patent application and searching for those terms in a database containing the full text of issued patents, patent publications, and publications qualifying as prior art.[54] Similarly, prior art in the form of use or other public availability can be searched for in databases likely to contain reports of such activity, press releases, advertisements, and other announcements.[55]
In addition to novel, an invention must be nonobvious under section 103.[56] Obviousness under section 103 results when one or more prior art references collectively contain all of the elements of the claimed invention and something motivates a person skilled in the relevant art to combine those references to come up with the invention.[57] The Supreme Court, in Graham v. John Deere Company,[58] offered what the Patent Office understood to be a relatively simple formula; in addition to the elements of the claimed invention, prior art had to contain some teaching, suggestion, or motivation (“TSM”) to combine them.[59] The John Deere Court also accepted the idea that “secondary considerations” could play a role in assessing obviousness:
Such secondary considerations as commercial success, long felt but unsolved needs, failure of others, etc., might be utilized to give light to the circumstances surrounding the origin of the subject matter sought to be patented. As indicia of obviousness or nonobviousness, these inquiries may have relevancy.[60]
That TSM test was applied as the core analytical principle of obviousness analysis until the Supreme Court decided KSR International Company v. Teleflex Inc. (“KSR”)[61] in 2007. In KSR, the Supreme Court rebuked the United States Court of Appeals for the Federal Circuit, the appeals court with exclusive jurisdiction over patent appeals, for applying too simplistic and rigid a test and held that a more flexible assessment of obviousness was required.[62]
USPTO synthesized KSR as requiring examiners to consider:
(A) Combining prior art elements according to known methods to yield predictable results;
(B) Simple substitution of one known element for another to obtain predictable results;
(C) Use of known technique to improve similar devices (methods, or products) in the same way;
(D) Applying a known technique to a known device (method, or product) ready for improvement to yield predictable results;
(E) “Obvious to try” – choosing from a finite number of identified, predictable solutions, with a reasonable expectation of success;
(F) Known work in one field of endeavor may prompt variations of it for use in either the same field or a different one based on design incentives or other market forces if the variations are predictable to one of ordinary skill in the art;
(G) Some teaching, suggestion, or motivation in the prior art that would have led one of ordinary skill to modify the prior art reference or to combine prior art reference teachings to arrive at the claimed invention.[63]
The consistent theme in all of the examples is whether a person having ordinary skill in the art (“PHOSITA”)[64] would have a “reasonable expectation of success,” in combining elements from the prior art references.[65] If a PHOSITA would not have a reasonable expectation of success in combining elements of different sources, the invention is not obvious. This reasonable-expectation-of success-standard is similar in many respects to the without-undue-experimentation test for enablement under section 112.[66]
An application for a patent, even for a novel and non-obvious invention, must be accompanied by a specification to be published as part of the patent. The specification must be sufficiently clear and definite that one can know what the inventor is claiming as his invention and sufficiently detailed and concrete to enable someone skilled in the art to make and use the invention with little more than reading the specification.[67]
Section 112 is interpreted to impose two requirements: first, that the specification “particularly point[] out and distinctly claim[] the subject matter which the inventor or a joint inventor regards as the invention,”[68] and that it describe “the manner and process of making and using it, in such full, clear, concise, and exact terms as to enable [one] skilled in the art” to make and use the invention.[69] The first function is largely performed by the claims section of the specification, required to be separated and labeled as such.[70]
Section 112 requires that the applicant make the metes and bounds of a claimed invention clear.[71] That requirement is performed by the claims, but is not limited to the claims; the more discursive description is intended to flesh out an understanding of the claims of the patent.[72] The enablement requirement of section 112 is exported into section 102 in section 103 analysis: a prior art reference is not anticipatory under section 102 unless it is enabling,[73] permitting person skilled in the relevant art to build and use an invention without undue experimentation,[74] and a combination of prior art references do not make an invention obvious unless they are collectively enabling.[75]
Patents are issued to “inventors.”[76] The inventorship concept was refined in U.S. patent law during its first 223 years, when priority was determined based on when invention occurred.[77] Priority disputes turned on who was the first to invent, and resolving those disputes necessitated rules for deciding when an invention was complete.[78] Was conception enough, or must there be reduction to practice? And what constitutes reduction to practice?
Those refinements still apply, now that the America Invents Act (“AIA”) makes priority depend on when an inventor files an application for a patent.[79]
The two key events in the birth of an invention are conception and reduction to practice. The distinction between conception and reduction to practice emerged early in patent cases. In Perry v. Cornell,[80] an inventor claimed that a competing inventor, even if he did conceive of an invention first, did not qualify as an inventor because “it was merely an intellectual invention, based on theory, and not an invention in the meaning of the law.”[81] The court rejected the argument. “There is no law requiring the applicant to reduce his invention to actual use before he can obtain a patent.”[82]
The statute defines how USPTO reviews applications and interacts with applicants,[83] and the Patent Office has, by rule, filled in the details[84] and requires use of a sophisticated digital processing system for handling claims in the interactive process.[85] Decisions by examiners can be reviewed in an administrative appeal by the Patent Trial and Appeals Board (“PTAB”).[86]
Patents last twenty years from the date of the first application,[87] extendable to compensate for delays caused by the Patent Office.[88]
After patents are issued, they may be revised by being withdrawn and re-issued,[89] or in ex parte re-examinations triggered by the patent owner or a third party.[90] Since 2021, they also can be challenged in administrative litigation before PTAB in Post Grant Review on almost any grounds for nine months after issuance,[91] and after nine months, in Inter Partes Review[92] when a challenger can marshal patents or printed publications[93] that establish a “reasonable likelihood that the petitioner would prevail with respect to at least [one] of the claims challenged in the petition.”[94] The validity of issued patents also can be reviewed by United States district courts adjudicating actions brought by patent holders for infringement[95] or in declaratory judgment actions brought by potential infringers.[96]
The system in the United States is integrated with other systems around the world by providing for international filings and recognition of filings in one country by the patent offices in other countries.[97]
“Although ‘AI’ is nearly ubiquitous, it has no agreed upon definition.”[98] Little doubt exists, however, that artificial intelligence is a form of computer programming.[99] Like other programming, it relies on abstraction,[100] representing real-world things and their behaviors by algorithms. Its hardware “understands” information represented as data organized in structures[101] that associate values with attributes of things and characteristics of behaviors. Its software makes use of iteration[102] to perform repetitive calculations rapidly and infers conclusions by applying algorithms[103] to appropriate data. It accomplishes these things and much more by stringing together simple operations like “fetch this value from memory,” “compare it with that one,” and “if they are the same, add them and put them somewhere else in memory.”[104] These characteristics of computer programming had crystallized reasonably well by the late 1930s and were the principal tools of computer programmers by the time the first digital computers became useful during and after the Second World War.[105] Computer science theory developed around these practices in the 1950s and 1960s.[106]
Digital computers sped up the computation of statistics according to formulas developed in the nineteenth century, and facilitated tagging, sorting, and comparing information artifacts—operations at the heart of ChatGPT.
The term “artificial intelligence” first became popular right after the Second World War when the popular press fueled excitement over the new “electronic brains.”[107] British cryptographer Alan Turing popularized the still-used test for intelligence: whether a human being interacting through an opaque communication mechanism can tell whether what is on the other end is a human being or a machine.[108]
Artificial intelligence enjoyed another upsurge in popularity in the 1980s. Experts in various fields trained computers to act as “expert systems,”[109] working with knowledge engineers, specialists skilled in human processing of information, to attempt to get computers to mimic them.[110]
Most recently, the press and policymakers have excited each other over artificial intelligence once again after an extremely successful public relations and marketing campaign by Sam Altman,[111] the CEO of OpenAI, an enterprise funded largely by Microsoft.[112] The natural language processing capability of his ChatGPT,[113] a publicly released, free generative AI system astounded reporters and others with its fluency in understanding and responding to natural language queries and its ability to generate relevant responses from machine learning applied to the enormous store of information available through the Internet. Celebrated for its glibness,[114] the generative AI of the 2020s represents merely incremental improvements in computer processing techniques that have been rapidly evolving over the past seventy-five years. In Ex Parte Daniel J. Ferranti, the Board rejected a claim limitation of “natural language processing (NLP) algorithms” because they are “well-known, conventional, and routine,” going back into the 1950s.[115]
Claims of doom because computers have suddenly become intelligent and willful[116] are based on ignorance of how generative AI works, based on marketing efforts by vendors, or reflect efforts by policy makers and commentators to get attention.[117]
Although the power of generative AI to upend society is greatly exaggerated, the capabilities of the technology mesh well with the nature of the patent system. The technology excels at understanding verbal queries and instructions. It has a powerful ability to extract and organize ideas and expressions from knowledge verbally expressed and recorded and accessible through the Internet and other digital means. It deals most comfortably with structured information formats.
All of these characteristics are present in the patent system. The full text of millions of patents and patent applications is available for free.[118] The questions and instructions relevant to patent applications and their prosecution are predictable. The formats of communications between applicants and the Patent Office and of patents themselves are settled and highly stylized. A good fit exists between generative AI and patent practice.
The most relevant AI technology involves natural language processing and large language models developed through machine learning.[119] Of secondary importance, but helpful in understanding how the technology works, is image processing. Generative AI also can write music, by predicting sequences of notes, chords, and rhythmic patterns according to musical works recorded in the past.[120] But musical AI is of less relevance to the subject of the article, except, perhaps, when an inventor of a music composition AI program seeks a patent. Generative AI and its predictive capabilities can be applied to music: in Western twelve-tone tonal music is a G more likely to be followed by a C or a C-sharp?[121] Is an imperfect cadence or deceptive cadence likely to be followed by a perfect cadence?[122] Is a rhythmic pattern likely to be repeated in the next measure?[123] The most visible of the recent advances involves pattern matching applied by cheap and powerful new semiconductor chips to the huge stock of information available through the Internet. Machine learning in large language models enables the systems to develop algorithms that predict what comes next in information artifacts.[124] At the lowest level, these algorithms predict what characters are likely to come next in the words of a particular language, what words are likely to come next in a grammatically correct sentence, what ideas (semantics) are likely to be associated with particular strings of words, and what ideas are likely to be associated with each other.[125] The “machine-learning” label applies to the process of identifying the relevant features, as many as have statistical significance.[126] There is nothing magical about the analysis: it is factor analysis,[127] which has been used as a social science methodology for more than 100 years.[128]
Machine learning trains large language models on large quantities of text data, and acquires knowledge of vocabulary, grammar, and semantics by analyzing and developing statistical rules of language.[129] Training data for large language models may come from a variety of sources, including Wikipedia, news articles, electronic books, and web page content on the Internet.[130] Through this enormous data store, a large language model can learn knowledge in different fields, so that the model is capable of wide application.[131] “Once the large language model is trained, it can be used for a variety of natural language processing tasks, such as text generation, machine translation, sentiment analysis, question answering systems” and the like.[132]
Large language machine learning models work by breaking words in a sentence into tokens, taking nearby words into account, aggregating these clusters of neighboring tokens into an enormous database, and using statistical methods to construct a vector (a list of values) representing each word’s relationship to other words and its connection with concepts.[133] It constantly adjusts the vector as its learning progresses.[134] Before transformers, the state of the art in natural language processing was a recurrent neural network (“RNN”), which processed each word in a sentence sequentially.[135] Transformers take entire sentences and “remember” the relationship of each word with all the others. Transformers allow models to take context from beyond sentence boundaries.[136]
Models of semantic relationships are developed by interactive comparison of sentences separately deconstructed by neural networks and computing a loss function that represents their similarity or lack thereof.[137]
It is a bottom-up learning process: first distinguishing individual words, then evaluating the frequency with which words appear together in phrases, then associating the phrases with concepts, and then building a hierarchy of concepts, a semantic tree,[138] not unlike a conventional thesaurus.[139] A loss function reflects how well the network does in predicting. Sophisticated networks feed back their results to compare them with actual data, permitting loss functions to be computed and then improved.[140]
A recent patent for generative AI[141] explains in some detail how machine learning is used to train a system for natural language output in a style that “write[s] like me.”[142] At a fundamental level, natural language analyzers and generators use an ontological network: a sophisticated, computerized thesaurus, which classifies linguistic concepts and organizes them according to their relationship with each other.[143] The result is a semantic framework for a particular language. Particular words are slotted into their appropriate ontological classes, thus organizing the entire vocabulary of the language.[144]
The system parses training text into sentences,[145] uses pattern matching to classify concepts expressed in each sentence and then assigns semantic tokens accordingly.[146] Deictic[147] context is developed using anchor words, which signify that the surrounding syntactical units should be processed as teaching examples. Examples of anchor words signifying comparison include “increase[,] reduction[,] decrease[,] decline[,] rise[,] fall[,] raise[, and] lower[.]”[148] Anchor words are specified for each semantic concept. A complete system may use scores of separately patented methods.[149]
The most sophisticated machine learning systems employ transformers—the “T” in ChatGPT.[150] Transformers are a type of neural network architecture that “remembers” what came before, maybe long before, in a sequence of words, ideas, sounds, or images.[151] They enable a better and more efficient prediction of “what comes next” than the common alternative, recurrent neural networks.[152] Transformers use convolutional neural networks combined with attention models.[153]
The performance of machine learning systems based on neural networks can be improved by constructing attention vectors, which rate the importance of each word in a cluster of semantically related sentences.[154]
Natural language user queries can be preprocessed to chunk a set of sentences of the natural language user query into a set of smaller sentences while retaining the references between chunks of the set of sentences. For each chunk of preprocessed user query a Name Entity Recognition (“NER”) ensemble extracts a domain-specific name entity from the chunked preprocessed user query.[155]
Classification techniques model the topics of each chunk. Then, in appropriate applications, sentiment analysis can determine a sentiment of each chunk of the user query.[156] User queries are converted to system queries by a combination of different natural-language-processing functions, such as NER, sentiment analysis, part of speech tagging, canonicalization,[157] classification, and translation.
Then, pattern matching techniques are used to compare the system query to the knowledge model to determine a closest state in the knowledge model.[158] The system returns a set of decisions scored according to their degree of match with the system query.[159] A winner state in the knowledge model is the knowledge state with the highest score.[160]
Systems can be fine-tuned by taking large language models trained on the full array of data usually used for training such models and then connecting them with more specialized learning databases.[161] The machine learning system constructs new concept vectors enabling a branching by subject matter before more finely grained responses are constructed.[162]
These predictions are organized in a hierarchy through neural networks using transformers.[163]
A natural language processing system using a large language model makes statistically robust predictions to construct words, then sentences, then to associate sentences with concepts, and to put the concepts together logically according to the way they have occurred in the store of information recorded in the past.[164] Image recognition and creation systems assemble pixels coherently into visual objects, which can be associated with animals or inanimate things like trees or rocks, and recreate movement by predicting where an object will be from one frame of a video recording to the next.[165]
Generative AI and machine learning work with images similarly, although the raw material is digitized images rather than text.[166] Images can be processed by predicting whether the next pixel to the right, scanning horizontally, is likely to be lighter or darker, the same or a different color, and if a different color, which.[167] The techniques work by scanning the lines of an image, much as a laser printer or office scanner does, and looking for discontinuities in brightness and color.[168] A model of an image then is constructed identifying the locations of those discontinuities.[169] Then, statistical algorithms, enabled by a complex hierarchy of neural networks,[170] compare the location of different types of discontinuities between images, and thus identify images that are most similar, and then, at a higher level, whether a lip is likely to be followed by a cheek or an ear, or a steer’s nose by a mouth or a tail.[171] A very large number of image samples are processed by machine learning.[172] Some of the samples contain the target image, and others contain something else.[173] Thus, a robocowboy[174] might be trained to recognize cattle by presenting hundreds of thousands of images of different kinds of animals, tagging only those that represent cows, bulls, steers, and calves.[175] The indicia of similarity are the particular facial features that discriminate a cow from a wolf—or one face from another.[176]
Machine-learning techniques can be used to accommodate the challenges associated with recognizing the target image despite different orientations, different lighting conditions, and different backgrounds.[177]
The techniques were developed for image matching applications such as face-recognition.[178] But the models developed through machine learning can be used in generative AI to create new images. The most important features for facial representation are “the gap between the eyes, the width of the noses, the length of the nose, the height and shape of the cheekbones, the width of the chin, the height of the forehead and other parameters.”[179] The feature measurements can be expressed as a “feature vector,” or “faceprint,” which represents a particular face.[180]
The ability to predict used in both large language models and in image processing is called generative, referring to the capacity of the systems to generate new patterns of expression.[181] Generative AI can respond to complex natural language prompts running to several paragraphs or many pages. It can capture the subject matter inquired about, and then generate sentences, paragraphs, or pages of relevant results.[182] They can summarize data, synthesize reports and other information artifacts such as statutes, judicial opinions, and patents.[183] They can accept instructions and generate visual portrayals, synthesizing photographs or caricatures. They can write poetry, some of it evocative, and music, some of it entertaining.[184]
What they cannot do is think. They cannot create beyond the boundaries of what has been created in the past. They are incapable of the flashes of genius that sometimes are thought to be characteristic of good inventors.[185]
They are, however, very good at routine clerical and office tasks. Just as USPTO was able to use the computer programming of the mid-1980s to move files around between applicants and examiners, so it can use the computer programming of 2025 to automate many of the clerical task inherent in implementing a highly stylized document review process.[186] Already, USPTO uses what most people would call artificial intelligence to screen electronic applications to make sure they have all their parts, formatted in the proper ways.[187] It uses full-text and semantic search programs to search for prior art.[188] These have been in use by Westlaw and Lexis, and by Google and other search engines for nearly thirty years, and are now quite sophisticated in accepting free text queries.[189]
Patent practitioners routinely use patent prosecution assistants, advertised as embodying AI, to draft patent applications and to vet them to assure consistency and appropriate internal references necessary for antecedent bases for claims language.[190] They also use specialized prior art search engines building on the techniques developed by Westlaw, Lexis, and Google.[191]
Some of the same tools can compare patent applications against prior art to flag anticipation and obviousness or can check specifications for compliance with section 112. They already spit out form objections and rejections by patent examiners and draft responses to those by patent practitioners.[192]
They can review specifications, drawing labels, and the language of claims already drafted to suggest additional claim language in patent applications. The same techniques that enable obviousness analysis also can be used to suggest new inventions.
How widely these capabilities will be adopted depends, not only on their capabilities in a day-to-day practice environment, but also on cost. The cost of these systems is just being encountered. Huge databases of information and enormous computing power are required to accomplish the machine learning necessary for useful predictive algorithms. Not only is the hardware expensive and available from only a few vendors, it uses large amounts of electricity.[193]
It is also not yet completely clear how sophisticated prompts can be handled in large language models without compromising confidentiality.
The story at the beginning of this article[194] touches on the principal ways that generative AI can be used in the patent system: as an inventor, as the subject of an invention, to prosecute applications, to perform examiner functions and to aid pro-se applicants. USPTO issued two policy statements on generative AI in 2023,[195] addressing the first and the third of these, intensifying public discussion already underway. Each of the following sections addresses one of these uses, explaining how the technology is already being deployed, analyzing legal issues presented by the use, and predicting how the technological capabilities are likely to be expanded.
Most of the excitement in the general press claims that a new era is dawning in which creative humans will be replaced by robots. Some of the loudest cries have come from authors and writers.[196] Authors’ claims that robots using AI would replace them was a theme of the 2023 writers’ strike against Hollywood.[197] One can imagine extending their alarm: AI will replace human inventors. This is nonsense. The clamor by authors and writers has little connection either with copyright law or with the way large language models learn.[198] It has more to do with the political power and organization of Hollywood writers, compared with AI engineers.
The February 2024 USPTO policy statement addressed the fantasy that AI inventors would replace human ones. Earlier, the United States Court of Appeals for the Federal Circuit had agreed with the Patent Office that only natural human beings can be inventors.[199]
In its policy statement USPTO considered the relationship between generative AI programs and their human masters in the process of inventing.[200] The statement reiterates the office’s position that only human beings can be inventors, not inanimate computer programs.[201] It then offers guidance on how human inventors working with AI partners can be inventors.[202] But human inventors can be aided by inanimate tools such as power drills, lathes, and milling machines, or, more recently by oscilloscopes, and digital computer simulations.[203] When such tools are involved, the question is a traditional one: does the human being qualify as an inventor—a question receiving much scrutiny over the 235-year life of the patent system? To qualify as an inventor, a person must contribute significantly to the conception behind the invention.[204] Developing an essential building block can qualify, so designing, training, or building a generative AI system backed up by machine learning can qualify one as an inventor.[205]
A natural person who only presents a problem to an AI system may not be a proper inventor or joint inventor of an invention identified from the output of the AI system. However, a significant contribution could be shown by the way the person constructs the prompt in view of a specific problem to elicit a particular solution from the AI system.[206]
On the other hand, “[m]erely recognizing a problem or having a general goal or research plan” is not enough.[207] Neither is reducing a conception developed by someone else or something else to practice.[208] And “maintaining intellectual domination” is not enough.[209]
The policy statement has a section specifically focusing on design and plant patent applications,[210] suggesting that the Office thinks those subject areas are most likely to spawn inventions with significant AI contributions and the most human reliance.
The policy statement emphasizes the duty of disclosure imposed on applicants and their coworkers with regard to determination of inventorship.[211]
Several witnesses in Senate Judiciary Committee hearings in 2023 told Congress to amend the patent statute to allow generative AI systems to be inventors, while other witnesses disagreed.[212] Even the witnesses favoring the possibility of inventorship for AI computer systems recommended that a human being be the applicant.[213] That position is a bit schizophrenic, because the typical applicant is a corporation, having been assigned rights by the human inventors. Accountability is focused on the applicant and the patent owner, which already can be—and usually is—an inanimate corporate entity. No one has been terribly clear or persuasive about what is gained by insisting that a natural person be named as an inventor, increasingly a purely symbolic act.
Ryan Abbott, Professor of Law and Health Sciences at the University of Surrey School of Law proposed at a hearing: first, that AI should be defined functionally for purposes of regulatory efforts and regulated in a technologically neutral manner.[214] Second, that the Patent Act should be amended so that AI-generated inventions are patentable, and so that patentability shall not be denied based on how an invention is discovered.[215] Third, that in the case of an AI-generated invention lacking a traditional human inventor, the AI system that has functionally invented should be listed as the inventor and the AI’s owner should be the owner of any intellectual property generated by their system.[216]
One of the difficulties with USPTO’s current position, he said, is that the human shepherd of AI systems may not qualify as inventors.[217] Not allowing AI to be an inventor renders an entire class of otherwise patentable subject matter unpatentable. This will get worse as AI continues to improve.[218] Without patentability, owners and users of AI systems will tend to protect their systems with trade secrets rather than patents.[219]
Corey Salsberg, vice president at Novartis,[220] argued that AI is not capable of inventing but that its use may cause examiners to deny inventorship status to human users.[221]
Laura Sheridan, Head of Patent Policy for Google[222] saw no necessity for allowing AI to be named as an inventor, because “in Google’s extensive experience using AI as a tool for innovation, humans are involved in a way that makes them inventors for the resulting innovations.”[223]
John Villasenor, UCLA Professor of Law and Electrical Engineering[224] testified “that AI inventions should be patentable, and that inventorship should be attributed to the natural persons who use AI as a tool to enhance their ability to innovate.”[225] This position, he thought, is fully consistent with the current state of U.S. patent law.[226]
Rama Elluru, Senior Director for Society and Intellectual Property at the Special Competitive Studies Project[227] argued that “[i]t is critical that AI-generated inventions are patentable . . . allowing AI to be named as an inventor on patent applications might be one way to recognize the patentability of AI-generated inventions.”[228]
It also is not clear that any statutory amendment is necessary to allow AI computer systems to be inventors. Nowhere in the statute is there an explicit requirement for human inventorship.[229] To be sure, throughout the life of U.S. patent statutes, inventors have been natural persons, but the idea that an inanimate business entity can be the applicant is fairly new to U.S. law, and the possibility of computer-generated inventive results meeting subject matter, novelty, and nonobviousness requirements is quite new to U.S. reality.
Consideration of USPTO’s guidance on inventorship and on the testimony spotlights the risks that more powerful generative AI systems used in the invention process may narrow the role of human users to the point that no one qualifies as an inventor under current USPTO guidance. Leaving a growing number of AI influenced inventions as orphans in the patent system would pull the system back from an important area of innovation. Critics of patents in general would be pleased with this, but it is not necessarily good policy.
Sean O’Connor observes that nothing in the Patent Act disables corporate persons from being inventors.[230] He finds no rational reasons why the “product of original . . . thought” aphorism used to mandate individual inventorship does not apply equally to copyrights, where the copyright act expressly allows corporate persons to be “authors.”[231]
He argues that both regimes should leave initial attribution of creation with natural persons, allowing them, however, to transfer control and ownership to corporate entities.[232] He justifies this outcome by his observation that attribution can be equally or more important than economic rights.[233] This is a plausible justification for preferring individual inventorship, but not one strong enough to deny legal accommodation to the reality of robotic invention.
Section 101 analysis would be considerably simplified, if the Patent Office did not care whether a named inventor is a natural human being or inanimate. Matters contained in oaths and declarations can be handled by the applicant or some other natural legal person taking responsibility for the inventive computer. Indeed the disclosure requirements under present law are not limited to natural-person inventors.[234]
Resolving the inventorship question in this way would not bias other aspects of eligibility under section 101. Proponents of easing eligibility for AI -related inventions can continue to make their arguments and continue to be opposed by those defending the status quo or urging further tightening.
The case against AI ownership of patents is weak. To be sure, computer systems running AI software on specialized AI hardware are not typically separate business entities, but they are owned by business entities.[235] Just as a business entity can be an applicant for a patent, so also can it be listed as the inventor. Matters such as oaths and declarations can be handled easily by requiring them to be executed by those business-entity employees most closely connected with the inventive activities of the AI system.
The fictional Edison is not going to get a patent on the application it drafted,[236] even if it is not disqualified because it is not human. The language does not come anywhere close to meeting the requirements of section 112,[237] and nothing in the session assessed novelty[238] or obviousness.[239] Generative AI systems are not going to replace human inventors, but some technological contexts exist in which a generative AI will do so much of the work of describing an invention and articulating claims. The law should not require applicants for patents in these contexts to turn cartwheels to emphasize the human contribution and to minimize the machine’s.
The Patent Office regularly is granting patents on AI inventions and receiving more applications. An October 17, 2024, advanced search in USPTO’s ppubs database with the term “artificial intelligence” resulted in 117,057 records, representing issued patents.[240] The same search run against the published applications database resulted in 171,472 records.[241] An advanced search of the issued patents database with the term “generative AI” produced seventy-eight results, and the same search against the published applications database produced 370 results.[242]
New generative AI inventions qualify for patents if they meet the tests for patentability under section 101,[243] as interpreted in Alice/Mayo.[244] AI applications are computer programs, embedded in powerful computing hardware.[245] The complexity of the software and the power (and power consumption) of the hardware is singular and largely unprecedented.[246] But the legal tests for patent eligibility are more than two hundred years old.[247] The generative AI invention must involve a machine, process, item of manufacture or compound; it must be more than a mere algorithm or theory or phenomena of nature.[248] It helps assure patentability when an idea is expressed in detailed pseudocode or actual programming code for computers,[249] and it helps when the invention involves implementation of software on a physical piece of hardware.[250]
It is clear that prompts for an AI system can satisfy the conception requirement.[251] The Office published examples to be used with its 2019 guidance.[252] Example 39 pertains to an AI invention.[253]
The example contains a claim:
A computer-implemented method of training a neural network for facial detection comprising:
collecting a set of digital facial images from a database;
applying one or more transformations to each digital facial image including mirroring, rotating, smoothing, or contrast reduction to create a modified set of digital facial images;
creating a first training set comprising the collected set of digital facial images, the modified set of digital facial images, and a set of digital non-facial images;
training the neural network in a first stage using the first training set;
creating a second training set for a second stage of training comprising the first training set and digital non-facial images that are incorrectly detected as facial images after the first stage of training; and
training the neural network in a second stage using the second training set.[254]
The example concludes that the claim “recites a series of steps and, therefore, is a process,” a statutory category.[255] It concludes that no judicial exception is recited because:
The claim does not recite any mathematical relationships, formulas, or calculations. While some of the limitations may be based on mathematical concepts, the mathematical concepts are not recited in the claims. Further, the claim does not recite a mental process because the steps are not practically performed in the human mind. Finally, the claim does not recite any method of organizing human activity such as a fundamental economic concept or managing interactions between people. Thus, the claim is eligible because it does not recite a judicial exception.[256]
The USPTO example does not offer any help on application of Step II of Alice, where most of the recent AI applications have run aground. Jon Grossman, of Blank Rome LLP, offers more guidance on that subject.[257] He reviews Federal Circuit, district court, and PTAB decisions, most of them finding AI inventions ineligible under section 101, and concludes that:
[I]t is not the terms used in AI technology itself but the meaning and scope of those terms in the claims and specifications that govern the result of the Alice test . . . . Second, it is important to look into the caselaw rationale regarding eligible AI subject matter to see a pattern that is applicable to your patent subject matter, and then follow the claims and specification examples associated with each pertinent case.[258]
He then makes four concrete suggestions:
A. Draft a claim that recites the specific function(s) or the improvement(s) explicitly tied to the AI features.
B. Draft a claim explicitly reciting the AI technology.
C. Draft a claim and a specification that do not merely improve the abstract idea of the claimed AI technology, but can directly associate the AI technique with improved hardware performance.
D. Draft a specification that discloses and supports hardware tied to the claimed AI technology not in terms of listing that hardware as generic components but as an improvement or a solution for a problem tied to the performance of such components.[259]
PTAB has considered appeals of rejections of inventions involving assertions of AI in two dozen cases reported since the beginning of 2024.[260] In only one, did the applicant survive section 101 scrutiny. In Ex parte Kristin E. McNeil, Robert C. Sizemore, David B. Werts, and Sterling R. Smith,[261] the Board, reversing the examiner, determined that claims were not directed to a judicial exception under Step IIA, Prong Two:
The Examiner’s determination that the testing and determining steps (steps [v] and [vi]) of claim 1 recite mental processes . . . is insufficient to support the rejection. The testing and determining steps (steps [v] and [vi]) of claim 1 only have meaning and exist in the context of a trained cognitive computing system, and thus, cannot be performed by human thought alone or by a human using pen and paper. . . . “Such problems only exist if there is a cognitive computing system that is trained through a machine learning process. Thus, the problem only exists in computer technology. Moreover, the solution provided by the invention is an improved computing tool that performs specific computer operations set forth in the claims to not only annotate inputs and outputs of a trained cognitive computing system, but to perform additional testing of the trained cognitive computing system using a test corpus and test questions and analysis of the results returned to thereby identify the source of the bias in the operation of the cognitive computing system”. . . . Thus, we determine the testing and determining steps (steps [v] and [vi]) of claim 1 are not merely part of the abstract idea; rather, they are additional steps performed in the context of a trained cognitive computing system that cannot be performed by human thought alone or by a human using pen and paper.[262]
The other cases involved a judicial exception and nothing more. In Ex Parte Corville O. Allen,[263] the board applied Step II, saying:
To do so, we look to whether the claim “appl[ies], rel[ies] on, or use[s] the judicial exception in a manner that imposes a meaningful limit on the judicial exception, such that the claim [is] more than a drafting effort designed to monopolize the judicial exception,” i.e., “integrates a judicial exception into a practical application.”[264]
It found that:
[T]he claim recites no technical or technological details on how the ingesting is performed such that not re-ingesting is a technical improvement or an improvement to the functioning of the computer. And to the extent the “expenditure of computer resources” is reduced, this reduction is a business improvement in terms of cost and use, and not a technical or technological improvement to the computer elements themselves.[265]
In Ex Parte Kevin Lyman, Li Yao, and Anthony Upton,[266] the Board similarly rejected a claim at Step II:
“[t]he claimed invention does not provide a technical improvement as the computing devices are used as a tool to implement the abstract idea.”. . . For example, the presence of additional hardware elements (e.g., “at least one processor,” “memory,” “client device” and “display device”) as well as AI as a training function in the claims are not indicative of “integration into a practical application.” Rather, these generic hardware elements and genericAI (i.e., any “supervised or unsupervised model”)are [sic] simply used as tools to verify billing and improper billing notification in a medical establishment. . . .The claimed generic computer components and generic AI are simply the “automation of the fundamental economic concept.”. . . “[M]erely require[ing] generic computer implementation,” “does not move into [§] 101 eligibility territory.”[267]
In Ex Parte Eduardo Cardoso, Alex Ferreira, Leonardo Santos, and Jaquelina Rocha,[268] the Board said more about the flaw in all of these rejected applications:
Taking the claim elements separately, the operation performed by the computer at each step of the process is expressed purely in terms of results, devoid of implementation details. Steps 1-4 recite generic computer processing expressed in terms of results desired by any and all possible means and so present no more than conceptual advice.
Steps 1-3 recite using an artificial intelligence component, but do not recite any technological implementation details as to how this is done or how the artificial intelligence component is constructed or operates. All recitations are no more than an instruction to apply a black box to labelled data inputs.
Limitation 5 recites that the artificial intelligence component is trained using historical employee off-boarding data to create a specialized machine learning model that determines the employee retention insights and the action steps for respective employers for particular employee types. This again simply recites applying some labelled data as input to a black box with no technological implementation details. The recitation of a specialized machine learning model that determines the employee retention insights and the action steps for respective employers for particular employee types is a conclusory description as a “specialized learning model” with no recitation of how it is so specialized or how the model is constructed or operates, followed by the aspirational objective for the model with no recitation as to how that objective is achieved.
Limitation 6 recites that the specialized machine learning model increases performance and accuracy regarding analytical and predictive capabilities of the artificial intelligence component thereby increasing performance of the computer, itself. This is likewise a conclusory recitation with no recitation of how the performance and accuracy are improved or how computer performance is improved.
All purported inventive aspects reside in how the data are interpreted and the results desired, and not in how the process physically enforces such a data interpretation or in how the processing technologically achieves those results.[269]
At least one of the PTAB cases rejected the application, not only under section 101, but also under section 112.[270] This shows how section 112 can properly be applied when a patent applicant makes a generic claim to artificial intelligence or to machine learning, especially since neither term is well defined in the art, let alone in popular understanding. It is impossible for anyone to know the scope of the claim. It is also impossible for someone skilled in the art to make and use the invention with no more than the specification in front of him, because he would have to invent artificial intelligence and machine learning, respectively. Moreover, black box limitations raise questions about whether the applicant had reduced the invention to practice—whether he “possessed” it. In Ex parte David Gustafson, Christopher Danson, Tomasz Stadnik, Jessica Leigh Hempel, Rachel Jean Stark, and Alain Stephan,[271] the Board explained that a “black box” limitation in a specification pertaining to machine learning does not satisfy the requirements of section 112 because it does not show that the inventors possessed the invention.[272]
The relationship between section 101 eligibility analysis and section 112 definiteness analysis has been controversial. In Berkheimer v. HP Inc.,[273] the Federal Circuit explained that “[t]he second step of the Alice test is satisfied when the claim limitations involve more than performance of well-understood, routine, [and] conventional activities previously known to the industry.”[274] It explained the application of that test by analogy to application of section 112.[275]
In April 2019, the Patent Office issued a memorandum explaining how the Office will apply Berkheimer.[276] It said that the analysis of whether an element is widely prevalent or in common use for purposes of applying Alice/Mayo Step II it the same as the analysis under section 112(a) as to whether “whether an element is so well-known that it need not be described in detail in the patent specification.”[277]
As Part III.C explains, Lee and Kayden have a lot of work ahead of them to refine the patent application drafted by Edison. If they do a good job of that, they may be able to get a patent on the CherryBot. Eligibility of the subject matter under section 101 depends on application of the same rules applied to any other computer system, as explained in Part II.A. Likewise, application of section 102’s novelty requirement and section 103’s nonobvious requirement will be straightforward, with patentability turning on their ability to show how the CherryBot is different from the prior art in ways that would not have been obvious to a PHOSITA.
How they apply section 112 is going to be the heavy lifting. The cases reviewed earlier in this section make it clear that Lee and Kayden are not going to be able to get a patent on Edison as a black box, and they will not help their cause by simply scattering buzzwords like “artificial intelligence,” “machine learning,” and “transformers” throughout their application. Edison could do that. They must explain how Edison works. That does not mean that they cannot adopt some off-the-shelf natural language engines and models, but they must explain how their use of those models differs from what has gone before in nonobvious ways, and they must do it with enough definiteness that a PHOSITA could follow the instructions represented by their specification and build their own Edison.[278]
Patent search and drafting of patent applications are good uses for generative AI. The subject matter is well defined, and the publicly available database of patents and patent applications is enormous, facilitating good machine learning with large language models used by Chat GPT and its competitors.
Google Patents offers full text search of prior art from around the world from one source. It provides pdfs of complete documents, images, citations and translations.[280] Its database includes more than 120 million patent publications from more than 100 patent offices around the world. It also offers access to technical documents and books indexed in Google Scholar and Google Books.[281] It offers instructions on use.[282]
Its natural language models are as powerful and flexible as Google Internet search as Westlaw and Lexis. Google Patents is more complete than USPTO PatSearch, and easier to use, as well. It is free.[283] It is a mainstay of prior art searching for both patent practitioners and patent examiners.
A half-dozen vendors are deploying generative AI for patent prosecution: Dolcera’s IP Author,[284] Rowan Patents,[285] Qatent,[286] DorothyAI,[287] davinci,[288] and XLScout.ai.[289] Of these, Rowan Patents and IP Author are the most useful. ChatGPT itself can play a useful role. Davinci requested that it be excluded from this review.
IP Author and Rowan Patents are ready for serious pilot projects in patent practices, and they have quite different strengths and weaknesses. IP Author’s integration with prior-art search capability is a considerable advantage, and it generates useful first drafts of all the elements of a patent application. Rowan Patents is a productive partner throughout the process of writing a patent application and offers assistance with prior art search.
Neither product can understand an invention; no AI technology can do that. Both can, however, tag phrases as concepts, and thereby relate them to other parts of the application.[290] Nothing within the foreseeable future will allow a computer program to examine a model or watch a demonstration and describe what’s inventive about it. So it always will be essential, as it is with all these products, for a user—inventor or patent lawyer—to provide some kind of starting point describing the invention—a brief disclosure, draft claims, or drawings.
IP Author can generate draft claims; indeed vanilla GPT4 can do that if supplied with a simple description. Rowan Patents is much more interactive than IP Author. IP Author is more of a black box. With any of the products, a skilled patent lawyer will have to pay close attention to claims language.
An ideal implementation would use IP Author to come up with an initial draft and then input that draft into Rowan Patents for managing drawings and the editing process.
Qatent is worth following, but it is not ready for any serious implementation. The publicly available version of ChatGPT does a good job of drafting “Field of the Invention” and “Background of the Invention” sections when given a relatively simply Abstract section as a prompt.
Both IP Author and Rowan Patents understand the structure of a patent application and organize pre-written material appropriately. IP Author concentrates on generating draft language through generative AI; Rowan Patents concentrates on managing terms and references according to MPEP standards. Provided with a brief description, IP Author drafted two independent claims and eighteen dependent claims in approximately 120 seconds. The user could edit the claims directly in a window to modify the invention, or to save the claims to draft a complete patent application. The software generated figures including a fairly detailed but primitive flowchart and a drafted a detailed description of the figures. The twenty-two page draft of the complete application could be downloaded in Microsoft Word.docx format.[291]
Working from the same brief description, IP Author displayed prior art search results. Each item of prior art had a graph rating similarity, along with brief paragraphs summarizing points of similarity and difference. IP Author prior art searches present search results in the form of a matrix, listing prior art references from the most similar to the least similar, and providing summary tables for each reference identifying the main similarities and main differences between the reference and the invention description used to prompt the search. The prior art identification and summary easily was downloadable in .docx format.
Dolcera’s IP Author’s work product was not perfect. It got the sequence of steps wrong in the flowchart. In another run, providing claims language along with a brief description confused it, triggering an endless loop. It produced claims language that interspersed system claims with method claims, repeated limitations in the independent claims, and expressed dependent claims in terms of the wrong independent claim. At one point, it hallucinated and produced drawings and associated detailed descriptions for an entirely different invention in a completely different field. Restarting the software fixed the problem.
Rowan Patents accepts claims, drawings, or brief descriptions as starting points. A user can begin by importing a Word file, in which case the system parses it and inserts portions into the appropriate sections of the Rowan patent. AI is just one small component of the software suite, which has been in use by the patent bar for a number of years. The program concentrates on defined data objects, such as claim terms, definitions, part numbers, and figure references. Once terms have been tagged, the system flags every instance of each term and provides a red warning for each term in the claims that does not also appear in the specification and in a drawing.
A user can enter text into defined windows, create drawings with provided tools, or import an existing .docx file by opening the entire file or by dragging and dropping components of it into the appropriate windows. A user can download a published patent or patent application, convert it to a .docx or to an .rp file, and cut and paste portions, editing them into a Rowan application.
One then can define new terms and parts by highlighting them and thereafter edit them. Edits propagate throughout the file. One can import an image, such as a drawing created by a commercial artist and apply numbers to it. One can import particular .pdf pages from a patent publication, crop them, and cover up existing elements such as numbers with new ones.
Rowan has a complete suite of drawing tools, including flow chart tools. It allows auto numbering using defined terms with point and click. It can auto generate flow charts and other figures. Using these tools, a user can generate a set of drawings more than adequate for initial submission to the Patent Office.
Initial implementation includes the ability to designate specific text in the draft patent application and then select from options of pre-built prompts for actions on that selected text. For example, selecting specific terminology provides options to define or describe that term. Selecting an entire paragraph (e.g., a claim) provides options to summarize that content. An optional bioscience module can keep track of DNA sequences.[292]
The product has limited prior art search features through its “launch analytics” feature. This feature breaks a completed patent application into pieces and sends the pieces back to Rowan servers. Then the product gives the user an option whether the user wants to see search results and flags items of prior art that might present 102, 101, 103, or 112 difficulties. It does not, unlike IP author, summarize similarities and differences. Rowan does not keep copies of an application submitted for its analytics module.
Rowant Patents is harder to use than IP Author because navigating around the different Rowan modules is not intuitive and the choices in each module are confusing. Two very detailed instruction manuals are available, however, which help when a new user gets stuck. As with any full-featured software product, use would become intuitive after a few hours’ experience.
Qatent generated reference list of thirty-one nouns and gerunds from submitted description and claims. The figures it generated were garbled and the flow chart was generic. The brief description of drawings was generic with things like “Figure 4 shows Block Chart 2.” The claims were repeated almost verbatim in the specification’s detailed description. Grammar and style were atrocious, with meaningless terms like triplication, performant, conductive, operative, and performative scattered throughout the description.
DorothyAI is only a search engine, with no drafting capability advertised. It is advertised as “directly compar[ing] the text of any document, such as inventions disclosures, patent claims, abstracts, product data sheets.”[293]
IP Author and Rowan Patents take secrecy and security seriously. Neither puts user provided information or software generated drafts into the Cloud, and every external exchange is encrypted.
Both the Dolcera and the Rowan products are ready for deployment as production elements in patent prosecution workflow—Dolcera’s IP Author at the threshold of a new application, Rowan Patents as a wingman throughout the application drafting process.
Generative AI technology is still trying to find a productive place in the legal world. Identifying useful work cases requires experienced practitioners to work with the technology developers and vendors to refine application concepts and product features.
The second USPTO policy statement addressed how AI technology can be used to assist inventors and their lawyers in writing and prosecuting patent applications.[294]
It began by noting the longstanding and ubiquitous use of computer tools in document drafting, including word processing software with spellcheck and grammar check.[295] More recently, the notice observes, word processing and other office-productivity software has adopted generative AI features that can develop a written document with much less human involvement.
For example, recent tools directed to the IP industry include the ability to draft technical specifications, generate responses to Office actions, write and respond to briefs, and even draft patent claims. The capabilities of these tools continue to grow, and there is no prohibition against using these computer tools in drafting documents for submission to the USPTO. Nor is there a general obligation to disclose.[296]
It then proceeded to identify a number of risks that practitioners must guard against when using intelligent computer tools. Practitioners must “take extra care” to verify the technical accuracy of specifications and drawings drafted by using AI tools and their compliance with the description and enablement requirements of section 112.[297]
Practitioners must ensure that readers can differentiate prophetic examples created or drafted by AI tools from actual working examples.[298] Practitioners are not relieved from ensuring that AI-written summaries of evidence, affidavits, petitions, and responses to Office actions, are accurate, both technically and legally. Such documents must “not introduce inaccurate statements and evidence into the record . . . .”[299] “[T]he signature requirement and corresponding certifications ensure that documents drafted with the assistance of AI systems have been reviewed by a person and that person believes everything in the document is true and not submitted for an improper purpose.”[300]
AI-conducted prior art searches must not “burden the Office with large numbers of cumulative and irrelevant submissions.”[301] Natural persons must sign Information Disclosure Statements (“IDS”s) certifying that they have “performed a reasonable inquiry—including not just reviewing the IDS form but reviewing each piece of prior art listed on the form—and determined the paper is compliant with 37 CFR 11.18(b).”[302] After review, irrelevant and marginally pertinent cumulative information must be removed.[303]
It notes confidentiality risks arising from use of AI systems. “Use of AI systems to perform prior art searches, application drafting, etc. may result in the inadvertent disclosure of client-sensitive or confidential information to third parties through the owners of these systems, causing harms to the client.”[304] When AI systems perform prior art searches or generate drafts of specification, claims, or responses to Office actions, they may retain the information on aspects of an invention. This information may be used to train AI models further or the data may be released to “third parties in breach of practitioners’ confidentiality obligations to their clients.”[305]
If Kayden submits the application for the CherryBot just as it was drafted by Edison, he will violate his duties to review the application and determine that it is factually correct and advances meritorious arguments. Hopefully he already knows that, but USPTO’s April 2024 guidance can serve as a reminder. Lee, if he signs an oath and declaration in support of the Edison-drafted application, will violate his duty as well. Lee risks having the patent subsequently found to be invalid because of his inequitable conduct in signing a false oath or declaration.[306] Kayden not only risks that; he risks disciplinary penalties under Patent Office rules of practice.[307]
Neither Kayden nor Lee has a duty to conduct a search for prior art.[308] But if they do, they must disclose anything they find that may be material to patentability.[309] An Information Disclosure Statement is the Patent Office’s preferred way to make that disclosure.[310]
They will do a search because they do not want to waste time and money on an application for an invention that is anticipated or obvious. They can get most of what they need through a combination of USPTO’s free Patent Public Search facility,[311] supplemented by a Google Patents[312] search. Google Patents has the advantage of including more material—foreign material, nonpatent publications—as well as providing instant translation of almost everything.[313]
If Kayden has access to Dolcera’s IP Author—surely Milligan, Anderson does—he and Lee will use its search engine[314] as well because of the good assessment of similarities and differences between the CherryBot and each prior art reference revealed by the search.
One former examiner and present patent agent described the tools available to examiners in 2020:[315]
The main patent search tool (EAST) is keyword-based and requires the Examiner to create a keyword description of the invention to search. Google Patents[316] and Google Scholar are included in many examiners’ favorite search tool set, especially for non-patent literature (NPL). The other artificial intelligence (AI) based search tools (e.g., ip.com) I used at the USPTO did not provide meaningful results when pasting in an entire claim the majority of time (maybe ~5% of time was accurate) but when they were right, they were very accurate. When I narrowed the search down to a handcrafted search string, I saw significant improvement in the AI based search tools.[317]
In February 2024, the Patent Office announced a five-year, $70 million, sole-source contract with Accenture to modernize internal patent office search tools.[318] The project emphasizes the use of AI and machine learning technologies “to expand, rank, and sort the results of existing patent search systems so the prior art that might otherwise not have been present in or near the top of a list of search results would not be readily available to examiners.”[319]
Nothing in the announcement or in the request for proposals or information statements that preceded the announcement suggests that the office is deploying AI beyond the search function. Nevertheless, the capabilities of existing private sector search tools like IP author show that search results can provide a considerable measure of analysis of similarity and differences between a patent application and a prior art reference, providing a significant starting point for novelty and obviousness analysis.[320]
As it does so it should disclose the resulting algorithms so they can be tested against long-decided interpretations of statutory requirements.
If the Office does not take the initiative to do this unilaterally, it should be subject to orders to make the disclosures under the Freedom of Information Act.
The procurement followed a Request for Information and Notice of Vendor Engagement published in August 2023, which provided details in a series of attachments.[321] The heart of the new system is a “More Like This Document” (“MLTD”) function which can take the full text of a patent application and identify other published U.S. patent documents, unpublished U.S. applications, and foreign patents from sixty-four countries.[322] MLTD “uses machine learning models and pre-computed measures of similarity.”[323] SimSearch sorts documents by degree of similarity. Users can “emphasize text or Cooperative Patent Classification (CPC) symbols within the patent application to tune the results from the model.”[324]
SimSearch includes the capability to suggest synonyms for an examiner to consider adding into a query. An examiner can select a single term, phrase, group of terms and phrases and receive contextualized groups of technical and general synonyms most similar to the examiner’s selection. The suggested synonyms are ranked by similarity to the examiner selection. Additionally, in order to make the results explainable, each of the returned synonyms shall include the specific word or phrase from within the examiner’s selection that identifies the strongest related synonym.
SimSearch includes the capability to suggest CPC symbols for an examiner to consider adding into a query. An examiner can select a single symbol, group of symbols, term, phrase, group of terms and phrases and receive CPC symbols most similar to the examiner’s selection. The suggested CPC symbols are ranked by similarity to the examiner selection. Additionally, in order to make the results explainable, each of the returned CPC symbols shall include the symbol, word, or phrase from within the examiner’s selection that identifies the strongest related CPC symbol.[325]
The Request identifies future enhancements, which could include capabilities that:
Leverage the Boolean queries based on keywords, patent classifications, and proximity operators as an “anchor” to retrieve and rank documents through AI models.
Direct image-to-image search by using examiner selected images from US or foreign patent documents or US patent applications as an “anchor” image to retrieve and rank documents through AI models.
Direct text-to-image search by using examiner Boolean queries or free text as an “anchor” to retrieve and rank documents through AI models.
Conduct chemical-based search by using chemical structures/formulas or chemical names as an “anchor” to retrieve and rank chemical structure, formula and chemical names and their corresponding documents through AI models.
Provide visualizations and indicators that provide users deeper insights into AI model reasoning, show users document similarity clusters and potential fields of search, and provides users with feedback about query effectiveness in yielding cited prior art references and documents tagged or dwelled upon.
Use free text and plain language (such as a description by the examiner of an invention or a technical concept) as a query mechanism for retrieval of documents in patent search.
Provide examiner access to an expansive Non-Patent Literature (NPL) corpus. In addition to the AI models that support a corpus of patent documents identified above, the USPTO seeks solutions with the ability for the AI models to support the incorporation of broad NPL sources. Solutions that can use examiner Boolean based queries, US or foreign patent documents, and/or US patent applications to crawl broad NPL libraries and publication sources, including behind paywalls, on the internet and return NPL citations or digital object identifiers in combination with snippets of publicly available content (such as abstracts or portions of documents freely accessible) for examiner review are preferred so that the examiner can leverage the Search tool to execute the search and USPTO paid subscriptions, document purchases, and/or databases of non-patent literature to obtain access to full copies. The ability for these NPL returns to be effectively scored by similarity in a manner that can be interleaved together with similarity rankings of patent documents are preferred.[326]
Experience with commercial prior art search products and USPTO’s request for similarity ranking shows that noise is an important factor limiting aggressive use of search engines. Being buried in prior art references is worse than having an incomplete set. With an incomplete set, one can begin section 102 and section 103 analysis. With a haystack of search results, one may not have time to find the needle.
The USPTO RFI does not suggest that the Office aspires to use machine learning to conduct anticipation or obvious analysis.[327] Anticipation analysis should be within reach because it involves little more than matching elements one-for-one.[328] A mechanical approach to first step in obviousness analysis—identifying references with some of the elements in the application[329]—is feasible now, but USPTO understands that a more nuanced, subjective, and holistic standard is required to assess motivation and expectation of success.[330] Such analysis is not within the current AI state of the art.
The patent bar can expect, indeed it should hope, that USPTO will continue to embrace new technologies for managing its enormous information load. As it does so, it should continue its policy of transparency on the tools it uses and the criteria those tools apply.[331] It is not unreasonable to suppose that the Office will use machine learning to develop algorithms to apply anticipation concepts under section 102’s novelty requirement, to apply the touchstones of obviousness under section 103’s nonobviousness requirement, and the test for adequate disclosure and enablement under section 112.
In mid-2024 USPTO solicited input on the impact of generative AI on prior art under the patent statutes. The Office published a request for information on April 30, 2024,[332] and held a public listening session on July 19, 2024.[333] The office posed a number of questions. The questions are surely relevant to the impact of generative AI on patent prosecution, but their expression and tone suggests an unwarranted defensiveness and reserve about a technology that has the potential greatly to improve patent office function. Responsive comments point out the advantages that new AI technology offer for patent prosecution and examination.[334]
Representation by a patent lawyer in patent prosecution is expensive. Some lawyers advertise bargain rates, but it is reasonable to expect that prosecuting a relatively routine application will cost upwards of $15,000, and if additional effort is required to overcome Patent Office resistance, the cost easily can climb towards $100,000.[335] Many inventors do not have the resources to pay those kinds of fees and therefore prosecute their patent applications pro se. They do not necessarily do an inferior job of it. Many of them have engineering or science credentials as strong as those required of patent agents, who are not required to be lawyers, but lack the same level of detailed knowledge about Patent Office procedures. Registered patent agents have passed the patent bar exam.[336] The typical inventor has not.[337] But now, AI software available at prices readily affordable by small inventors has much of the necessary knowledge about Patent Office procedure built in.
One study shows that pro-se applications were abandoned at a 76% rate compared with 35% of represented applications, because of pro-se applicants’ failure to understand USPTO requirements, because of unjustified early surrender, and because desired claims do not meet patent requirements and cost.[338] The data manipulations necessary to identify pro se applicants in the study suggest that USPTO does not identify applications filed pro se.[339] So time series data on pro-se applications necessary to know trends is not available.
It’s likely that the availability of patent prosecution software like that described in Part III.C will lead to an upsurge in pro se applications. At the margin, this may lead to a decline in business for its patent lawyers, although to some extent it will simply result in applications being filed that never would have come to a lawyer.
This possibility intensifies the challenge to law firms, which have been rethinking the viability of hourly billing for decades.[340] Law firms can respond by understanding their cost structures better so they can offer flat fee arrangements more frequently,[341] by using at least the same tools that are available to pro se applicants, using them more efficiently because they collectively understand better than the ordinary inventor how the new tools fit with Patent Office procedure, and, of course, because they have additional skills rooted in their knowledge of the law.
Generative AI will not result in particularly dramatic changes in the ways that trial lawyers use computer processing in patent infringement lawsuits. They already use sophisticated search tools to search the Patent Office database, Google patents, Google itself, and other search engines to detect instances of possible infringement.[342] They already use electronic discovery tools to process large quantities of emails and other documents in electronic form made available in civil discovery.[343] The advances in generative AI will improve both these types of tools but not result in any transformation of how they work or how they are used; artificial intelligence has been a core part of these tools for ten or fifteen years. Indeed, natural language processing was one of the original features that drew lawyers to Lexis and Westlaw in the early 1990s,[344] and it likewise was an original feature of Google and other Internet search engines in 1999.[345]
Based on the analysis in the preceding sections, Kayden will advise Lee that Edison will have difficulty getting named as an inventor, but that, under USPTO’s 2024 statement on inventorship, Lee should qualify as an inventor of the CherryBot, based on his prompts to Edison. He will tell Lee that they may be able to obtain a patent on Edison, if they claim the details about how Edison works and uses generative AI technology in the particular context of agricultural robotics. He will explain that both of them have legal obligations to review Edison’s draft patent application carefully and to revise it to comport with USPTO requirements for applications and to express what Lee believes is different from the prior art.
They will amuse themselves by speculating about the possibility of a robotic patent system, in which human inventors and their lawyers struggle to get human attention.
Beyond their (hypothetical) situation, the new generative AI technologies will produce two phenomena: a more robotic prosecution system and more pro-se applications and prosecutions. Observers who are concerned about these developments must marshal evidence and arguments about why they matter in a patent system that is falling so far short of meeting the needs of inventors that most of them prefer to rely on trade secrets or nimble entrepreneurship rather than patents.
Generative AI will intensify the trend for patent prosecution to involve the interaction of robots rather than of patent examiners and patent lawyers exercising discretion as to their positions and the expression and defense of them. Patent Center screening algorithms already check for format compliance and inclusion of all parts of an application. A relatively modest modification in the software would permit the same or similar software to assist in the examination process by performing an “every noun” check to ensure antecedent bases for all of the material elements in claims, as Rowan Patent does. Similarly, it is not difficult to catalog every limitation in the set of claims, as Rohan Patent does, for its compliance checking, and use the terms in that catalog as search terms in prior art searches aimed at identifying the main candidates for anticipation under section 102 novelty analysis.
Automating the obviousness analysis is harder. The author’s experiments with ChatGPT did not produce useful results. The robot just reiterated the legal requirements for nonobviousness without providing any concrete guidance as to whether a sample invention description was or was not obvious, as the following result shows:
Regarding your question about whether the described autonomous herding system is obvious under Section 103 of the United States patent law, this requires a legal analysis that involves the consideration of the state of prior art at the time the invention was made, whether there are any prior art references that are pertinent to the claims of the invention, and whether the differences between the prior art and the claims would have been obvious to a person having ordinary skill in the art (PHOSITA).
The autonomous herding system you described integrates advanced navigation, propulsion, sensory technology, optical-LiDAR sensors, global navigation satellite systems, and machine learning. This suggests a complex interplay of technologies tailored for a specific application—managing and retrieving stray animals—which could be argued to be non-obvious if each of these integrations provides a novel and unexpected benefit. The non-obviousness argument can be strengthened if this combination achieves more than the sum of its parts or solves a problem that prior systems could not solve.
In practice, a patent examiner would review relevant prior art, including existing patents, patent applications, and other published materials, to determine if the invention is non-obvious. He would consider whether the claimed invention is a predictable use of prior art technologies, or whether it reflects an inventive step that would not be obvious.
A patent attorney or agent would be better positioned to conduct this analysis and provide a legal opinion on the matter, as this would require a detailed investigation of the current state of technology, patent databases, and potential analogous fields. If you’re considering patenting this system, consulting with a patent professional would be a necessary step.[346]
Automating the section 112 analysis is even more challenging, but further experimentation with appropriate generative AI engines would be useful in quantifying the probabilities that this can be useful.
As generative AI technology improves, the resources available to the patent system will increasingly be diverted to computer resources and away from human resources. The result will be a system in which applicants and their lawyers interact with computers and have trouble obtaining real involvement by a human being with any kind of discretion to deal with exceptional cases. Some patent examiners already frustrate applicants in this way by pumping out standard forms and not seeming to pay any attention to applicant arguments.
This problem is already familiar to anyone who engages in e-commerce and interacts with call centers. Vendors make it increasingly difficult to escape customer support corrals on websites and get to human beings. When a human being is available, he or she all too often is completely straitjacketed by what he or she sees on a computer screen.
The second phenomenon likely to result from the spread of generative AI technology in the patent system is that the technology will reduce transaction costs significantly, and that will result in more patent applications and may result in a higher proportion of applications being turned into patents. Even if the acceptance rate does not increase, the number of patents is linearly related to the number of applications.
Part of this will be an upsurge in pro se patent applications, which are also being encouraged by well-funded patent office initiatives to educate and assist pro se applicants. These initiatives are mandated statutorily in the interest of increasing racial, gender, ethnic, and wealth diversity in the patent system.
Regardless of how interesting these phenomena are, a serious student of law and public policy should ask “Who cares?” Most patents never are commercialized; let alone do they become subjects of infringement controversy. Many more pro-se applicants are abandoned than represented applications.[347] A significant increase in the number of patents will almost inevitably result not only in a greater number but also a greater proportion of patents that don’t make any real difference to commercial reality. This result would not necessarily be harmful, except that it will trivialize patent law and the significance of owning a patent. Trivialization generally is not a good thing for the legitimacy of law.
Fifty years, or so, ago Quaker Oats, as a marketing ploy, purported to give fee simple interests in one square inch parcels of land in Alaska.[348] It is not clear how legally effective these promises were, but some commentators have thought about its impact on Alaska, on the owners (which include the author, he thinks), and on property law, and have concluded that the overall effect was insignificant.[349] The data is clear: most patents are never commercialized.[350]
Moreover, “[m]any valuable inventions that are commercialized are not patented. Companies choose a variety of strategies to protect their inventions and intellectual property. For example, U.S. companies rate trade secrets higher than patents in their importance for protecting intellectual property[.]”[351]
[P]atent protection may be sought for reasons other than intended commercialization. Privately owned patents may be obtained to block rivals and negotiate with competitors, to use in lawsuits, or to build “thickets” of patents to impede or raise others’ costs of R&D and innovation . . . . New and emerging firms may seek patent protection to help obtain financing because investors perceive patents as potentially valuable for a firm’s assets and future profitability.[352]
More than half of issued patents are allowed to lapse before running their full twenty-year term.[353] In 2020, 62% of patent applications were abandoned.[354]
Not only may the impact of generative AI be uninteresting, though significant on the system, a number of barriers exist to wider utilization of generative AI, even in the patent arena, where its strengths seem particularly well matched to the needs. One constraint is the huge cost of the semiconductor chips necessary to do machine learning and operate large-language and other generative AI models.[355] The proponents of the technology have creatively and quite successfully sought to draw people into using their products, but they have done so with a considerable subsidy by making it free or almost free to everyone. At some point, they will have to increase prices in order to cover their costs and provide a return on investment. The price elasticity of demand for generative AI is yet to be tested.
A second barrier is the enormous amounts of electricity required to operate computers engaged in machine learning and employing the results.[356] Already, NIMBY[357] activists are using energy consumption as a basis for opposing large data centers, which are being expanded in anticipation of a continued generative AI boom.[358]
The final constraint is the likelihood that the open range represented by Google and its search engine competitors will be fenced off, eradicating the business model that enables the machine learning for ChatGPT, in much the same way that long cattle drives were eradicated by barbed wire enclosures on the Plains one hundred fifty years ago.[359]
Most of the copyright claims by rights holders whose works have been included in Google like searches to feed machine learning have only frivolous claims of copyright infringement.[360] But the merit of their claims does not matter. Rent seeking of this sort is neither immoral nor, in this context, illegal. A new stream of revenue has sprung up originating in some general sense with the content the copyright owners have made available, and they want to share in the stream. That desire will continue regardless of whether they can succeed in any legal theory to compel sharing. Already, the owners of significant content that heretofore has been freely available on the Web are erecting pay walls around it—the twenty-first century equivalent of barbed wire.
But the patent arena is shielded from much of this controversy. Patents are public, as are published applications. A patentee has no claim of misappropriation or infringement against an AI machine learning system that learns from his patent.
The questions about how far AI will revolutionize invention, patent prosecution, and patent enforcement have more to do with economics than with policy and technology. AI’s penetration into the system already is substantial is and is likely to increase.
Whether a computer program can be an inventor is largely a distraction.
The Article assumes that the content of this Appendix represents the transcript of the session between the fictional Lee and the fictional AI robot Edison.[361]
[Henry Perritt]: what areas in robotics are most in need of new inventions
ChatGPT: Robotics is an evolving field that intersects with many areas of technology and human endeavor. As of my last update, here are some areas within robotics that are ripe for innovation:
1. Human-Robot Interaction (HRI): As robots become more prevalent in everyday settings, improving how humans and robots communicate and work together is crucial. This includes advancements in natural language processing, emotion recognition, and non-verbal communication cues.
2. Robotic Dexterity and Manipulation: Enabling robots to handle objects with the same dexterity as humans, especially in unstructured environments, is a significant challenge. This includes the development of advanced grippers and hands, as well as the algorithms that control them.
3. Autonomous Navigation: While significant strides have been made in this area, there’s still room for improvement in GPS-denied environments, uneven terrain, and dynamic, unstructured settings.
4. Power and Energy Efficiency: Longevity and energy efficiency in robots, especially for those that are mobile, are critical. This could include new battery technologies, energy harvesting, or more efficient motors and actuators.
5. Robot Ethics and Safety: As robots become more autonomous, ensuring they make decisions that are ethical and safe for all involved is increasingly important. This includes both the programming of ethical guidelines and the creation of safety measures to prevent accidents.
6. Swarm Robotics: Coordinating multiple robots to work together seamlessly can vastly improve efficiency in various tasks. Invention in this area would likely focus on communication methods and collective AI algorithms.
7. Soft Robotics: Robots that can change their stiffness and shape on demand could be incredibly useful in delicate environments or in situations where traditional robots would be too rigid or dangerous.
8. Biomimetic Robots: Robots that mimic the functions and movements of living organisms offer numerous potential applications, from new forms of locomotion to robots that can better adapt to natural environments.
9. Robotic Fabrication and Construction: Advancements in robots that can build or fabricate structures, especially in remote or hazardous environments, could change the landscape of construction and manufacturing.
10. Healthcare Robotics: There’s a continuous need for inventions in robotic surgery, rehabilitation robots, assistive devices for the elderly or disabled, and robotic prosthetics.
11. Robotic Materials: Materials that can self-heal, change properties in response to stimuli, or offer new ways to interface with electronics are needed to advance robot capabilities.
12. Environmental Monitoring and Clean-up: Robots that can monitor environmental health or clean up pollutants in areas where it’s unsafe or impractical for humans to do so.
13. Rescue Robots: Robots capable of navigating disaster sites to locate and rescue victims or deliver essential supplies can be vital during emergencies.
14. Space Robotics: The extreme conditions of space require robots that can operate with high reliability over long periods without maintenance, handling tasks like repair, construction, and exploration.
15. Agricultural Robotics: Inventions that could automate tasks like picking, weeding, or harvesting in ways that are gentle on the crops and efficient in varied farm conditions are in demand.
Each of these areas offers substantial opportunities for impactful inventions, driven by the need to solve complex problems and improve the capabilities and integration of robots in our lives. Collaboration between different scientific and engineering disciplines is often necessary to push the boundaries in these areas.
[Henry Perritt]: invent an automatic cherry picking robot
ChatGPT: Designing an automatic cherry picking robot involves considering various technical aspects to ensure the robot is efficient, gentle with the fruit, and adaptable to different cherry tree configurations. Here’s an outline for such an invention:
1. Automatic Cherry Picking Robot “CherryBot”
a) Design Overview
• Structure: Lightweight, with a central processing unit mounted on an all-terrain, stabilized platform suitable for navigating orchard rows.
• Mobility: Autonomous guided vehicle (AGV) system with electric-powered wheels, GPS navigation for orchard layout mapping, and obstacle detection sensors to navigate around trees and terrain.
• Power: Solar panels with backup batteries, allowing for extended periods of work with minimal environmental impact.
• Manipulator Arms: Multiple articulated arms equipped with soft, adaptive grippers that can gently pick cherries without bruising them.
• Sensors: A combination of visual, LiDAR, and infrared sensors to detect ripe cherries and assess the 3D structure of the trees for optimal picking paths.
b) Functional Modules
• Detection Module: Utilizes machine learning algorithms trained on a dataset of ripe cherries to accurately identify the target fruit. It adjusts picking strategy based on size, ripeness, and cluster density.
• Navigation Module: Processes input from GPS and onboard sensors to move the robot around the orchard and position it correctly in front of each cherry tree.
• Manipulation Module: Employs advanced control algorithms to guide the manipulator arms and grippers, ensuring precise and gentle picking of the cherries.
• Storage Module: Features a cushioned collection bin that minimizes impact and damage to cherries as they are gathered.
• Decision-making AI: Integrates sensor data and controls all systems to optimize the path of the robot and the picking process, learning from each interaction to improve efficiency.
c) Unique Features
• Noise and Vibration Control: Operates quietly to not disturb local wildlife or residents, with vibration-damping materials and technology to avoid shaking the trees excessively.
• Weatherproofing: Built to withstand various weather conditions with waterproof and temperature-resistant materials.
• Real-Time Monitoring: Equipped with a communication system that allows remote monitoring and control if manual intervention is required.
• Selective Harvesting: Capable of selectively picking fruit to meet specific quality or ripeness criteria, which can be adjusted via the operator interface.
• Collaborative Operation: Can work in tandem with other robots and human workers, coordinating activities to maximize efficiency.
d) Potential Development Stages
1. Prototype Development: Build a scaled-down version of the robot to test the picking mechanism and basic navigation.
2. Field Testing: Deploy the prototype in an actual orchard to gather data and refine the machine learning models and control systems.
3. Commercial Model Fabrication: With data and feedback from field testing, develop a full-scale, commercial-ready model.
4. Market Integration: Introduce the robot to the market with on-site demonstrations and pilot programs with early adopters.
5. Iterative Improvement: Continuously improve the design based on real-world performance data and customer feedback.
This automatic cherry picking robot could significantly increase harvesting efficiency and reduce labor costs, while also being gentle on the fruit and the environment. The technology used could potentially be adapted for other types of fruit-picking tasks as well, making it a versatile tool in the agricultural robot sector.
[Henry Perritt]: write a patent application for this invention
ChatGPT: Writing a complete patent application is a complex and detailed process usually undertaken by a patent attorney or agent, but I can give you an outline of what such an application might look like for the automatic cherry picking robot “CherryBot”.
Title of Invention: Autonomous Cherry Harvesting Robotic System (CherryBot)
Background of the Invention: The invention relates to agricultural robots, and more specifically to a robot capable of autonomously detecting, picking, and collecting cherries.
Field of the Invention: The invention falls within the field of robotics, particularly applied to precision agriculture and automated harvesting systems.
Summary of the Invention: This invention provides an automatic cherry picking robot designed to operate autonomously within cherry orchards. The robot is equipped with advanced navigation, sensory, and manipulation technologies to detect and harvest cherries.
Brief Description of the Drawings:
• Fig. 1: A side view of the CherryBot, illustrating the solar panels and the articulated manipulator arms.
• Fig. 2: A top-down layout of an orchard with the robot’s navigation path.
• Fig. 3: A close-up of the adaptive gripper handling a cherry.
• Fig. 4: Block diagram of the robot’s decision-making AI and control system.
Detailed Description: The CherryBot consists of several components that work synergistically to achieve efficient cherry picking.
1. Autonomous Navigation System: Description of the GPS and sensor-based navigation system.
2. Sensory Module: Details of the visual, LiDAR, and infrared sensors used to identify ripe cherries.
3. Manipulator Arms and Grippers: Explanation of the robotic arms, their range of motion, and the soft grippers designed to handle cherries without causing damage.
4. Machine Learning Detection System: A description of the algorithms used to identify cherries and determine the optimal picking strategy.
5. Collection and Storage System: Description of the mechanism that collects and stores cherries during operation.
6. AI and Control System: In-depth information about the AI that integrates sensor data, controls the robotic systems, and optimizes picking paths.
Claims: A series of numbered claims that define the scope of the invention, starting from broad to increasingly specific. For instance:
1. An automated cherry picking robot comprising:
• an autonomous guided vehicle chassis;
• a solar panel energy system;
• a plurality of articulated manipulator arms with adaptive grippers;
• a multi-sensor detection system integrating visual, LiDAR, and infrared sensors;
• a machine-learning algorithm trained to identify ripe cherries;
• an autonomous navigation system utilizing GPS;
• a collection bin for harvested cherries.
Further claims would break down these components into more specific elements and may claim the methods by which the robot picks cherries or navigates the orchard.
Abstract: A summary of the invention, highlighting the autonomous cherry picking robot with its advanced detection, navigation, and harvesting capabilities designed to work in orchard environments.
Conclusion, Ramifications, and Scope of Invention: A closing statement summarizing the innovation, potential variations, and the breadth of the invention’s applicability.
Please note, an actual patent application must also include professional drawings, a detailed description of each component and process, formal claims that precisely define the invention’s scope, and an abstract. The application would have to comply with all requirements of the patent office in which it is being filed. It is recommended to work with a patent attorney to ensure that the application meets all legal standards and to navigate the complex process of patent prosecution.
[Henry Perritt]: How long should the manipulator arms be?
ChatGPT: The length of the manipulator arms for an autonomous cherry picking robot like the CherryBot would need to be carefully considered based on several factors:
1. Tree Size: Typical cherry trees can vary in height and width. Dwarf varieties might be as short as 8-10 feet, while standard trees can reach heights of 35 feet or more. The robot arms should be long enough to reach at least halfway across the average tree canopy to minimize the need to reposition the robot.
2. Reach and Coverage: The arm length should be optimized for the maximum reach while ensuring stability and precision. It should allow the robot to cover a significant portion of the tree without moving too often, increasing efficiency.
3. Weight and Balance: Longer arms will be heavier and could affect the robot’s balance and power consumption. The design should ensure that the robot remains stable, even when the arms are fully extended.
4. Mobility Constraints: The arms should be able to retract or fold to allow the robot to navigate between rows of trees without damage to the trees or the robot.
5. Working Envelope: The concept of a working envelope in robotics refers to the range within which a robot arm can operate. The arms should have a working envelope that covers the necessary spatial area for effective cherry picking.
6. Safety and Practicality: The arms should not be so long that they pose a safety risk to humans or create practical issues with transport or navigation.
As an initial design consideration, if we assume the average mid-sized cherry tree is about 15-20 feet wide and the robot is positioned close to the trunk, manipulator arms of about 8-10 feet might be sufficient to access much of the tree. However, for larger trees or to minimize repositioning, arms may need to be longer. It’s also possible that a telescopic design could be implemented to provide variable length for different scenarios, adding to the complexity but also to the versatility of the CherryBot.
Final specifications should be determined after a thorough analysis of the target environment, engineering constraints, and after iterative testing and prototyping.
[Henry Perritt]: What algorithms can CherryBot use to identify cherries and determine the optimal picking strategy.
ChatGPT: To effectively identify cherries and determine the optimal picking strategy, CherryBot could utilize a combination of algorithms focused on image processing, machine learning, and path planning:
1. Image Processing Algorithms:
• Color Thresholding: To identify cherries based on their distinct red color against the green leaves.
• Edge Detection: To outline the boundaries of cherries and differentiate them from the background.
• Blob Detection: To locate the individual cherries or clusters of cherries as single entities.
2. Machine Learning Algorithms:
• Convolutional Neural Networks (CNNs): Trained on labeled datasets of cherry images to recognize the fruit in various stages of ripeness and under different lighting conditions.
• Object Detection Models: Such as YOLO (You Only Look Once) or SSD (Single Shot Multibox Detector), which can identify and locate cherries in real-time.
• Semantic Segmentation: To precisely segment the image into regions that represent different objects, separating cherries from leaves and branches.
3. Pattern Recognition:
• Shape Matching: To distinguish cherries based on their characteristic shape, especially useful in detecting partially obscured cherries.
4. Deep Learning for Ripeness Assessment:
• Reinforcement Learning: To continually improve the selection criteria based on the feedback from the actual ripeness of picked cherries versus the predicted ripeness.
5. Path Planning Algorithms:
• Greedy Algorithms: For initial implementations, to simply choose the closest ripe cherry next.
• Travelling Salesman Problem (TSP) Solvers: To find an efficient route that allows the robot to pick all identified ripe cherries with minimal movement.
• Dynamic Programming: To solve complex decision-making problems where the optimal solution involves solving smaller sub-problems.
6. Optimal Control Algorithms:
• Model Predictive Control (MPC): To control the manipulator arms by predicting the future state of the system and correcting the path in a controlled manner.
7. Safety and Collision Avoidance:
• Motion Planning Algorithms: Like Rapidly-exploring Random Trees (RRT) for planning the arm movements without colliding with the tree or other cherries.
8. Multi-Agent Coordination (if multiple robots are used):
• Swarm Intelligence Algorithms: To allow multiple robots to work together efficiently without interfering with each other.
9. Harvest Optimization:
• Genetic Algorithms: To optimize the picking schedule and pattern based on past performance and continuously evolving environmental variables.
All these algorithms would be part of a larger integrated system where the output of the image processing and machine learning algorithms informs the path planning and manipulation control systems to perform the actual picking of cherries. The system would be trained and tested using a variety of data from real-world scenarios to ensure robust performance under different conditions and to handle anomalies effectively. The combination of these algorithms would allow CherryBot to adaptively and intelligently harvest cherries, maximizing yield and efficiency while minimizing damage to the fruit.
[Henry Perritt]: What are the details of the CherryBot’s visual, LiDAR, and infrared sensors used to identify ripe cherries.
ChatGPT: For the CherryBot to efficiently identify ripe cherries, it would be outfitted with a suite of advanced sensors, each contributing unique capabilities to the overall system. Here are the details for each type of sensor:
2. Visual Sensors (Cameras)
1. High-Resolution Cameras:
• Purpose: Capture detailed images of the cherry trees, allowing for the identification of cherries based on color, shape, and size.
• Specifications: Multispectral imaging capabilities to distinguish ripe cherries from unripe ones and foliage.
• Integration: Real-time image processing with onboard GPUs for quick analysis.
2. Stereo Vision:
• Purpose: Provide depth perception to assess the distance and spatial arrangement of cherries.
• Specifications: A pair of cameras mimicking human binocular vision, calibrated for depth estimation.
• Integration: Coupled with computer vision algorithms to generate 3D models of the picking environment.
3. LiDAR Sensors
1. 3D LiDAR Scanners:
• Purpose: Generate detailed 3D maps of the orchard environment, including the structure of trees, branches, and the position of cherries.
• Specifications: High scan rate for rapid point cloud generation, capable of operating in bright sunlight and varying light conditions.
• Integration: Data fusion with visual sensors to enhance detection accuracy and to guide the manipulator arms precisely.
2. Miniaturized LiDAR:
• Purpose: Serve as a proximity sensor for the manipulator arms, avoiding collisions and fine-tuning the picking process.
• Specifications: Small form factor LiDARs placed on the robot’s arms to provide immediate spatial feedback.
• Integration: Real-time obstacle detection and avoidance during arm movement.
4. Infrared Sensors
1. Thermal Imaging Cameras:
• Purpose: Detect the temperature variations that can indicate ripeness and differentiate fruit from leaves.
• Specifications: Infrared cameras sensitive to the specific thermal signature of ripe cherries.
• Integration: Combined with visual data to improve detection accuracy, especially in low visibility conditions.
2. Near-Infrared (NIR) Spectroscopy:
• Purpose: Analyze the chemical composition of cherries to determine sugar content and ripeness.
• Specifications: Sensors capable of capturing NIR spectra that are indicative of different ripeness stages.
• Integration: Cross-referencing spectral data with visual and 3D LiDAR data for comprehensive ripeness analysis.
5. Combined Sensor System
• Sensor Fusion: All sensor data is integrated through sensor fusion algorithms, providing a comprehensive dataset that includes visual appearance, 3D structure, and thermal and chemical properties.
• Calibration System: Ensures that all sensors are calibrated to work in unison, providing accurate and consistent data.
• Data Processing: Utilizes edge computing with a combination of onboard processors and AI accelerators to process sensor data in real-time.
• Environmental Adaptation: Algorithms capable of adjusting sensor parameters based on environmental conditions, such as sunlight intensity, to maintain sensor performance.
• Feedback Loops: The sensors provide feedback to the control system, allowing for continuous adjustment and improvement of the picking process.
The details of these sensors represent a sophisticated system that, when combined with advanced algorithms, enables the CherryBot to navigate the orchard environment, identify ripe cherries, and execute the picking process with high precision and efficiency.
[Henry Perritt]: Please rewrite the patent application using these answers
ChatGPT:
Title of Invention: Autonomous Cherry Harvesting Robotic System (CherryBot)
Background of the Invention: The present invention pertains to agricultural robotics, specifically an autonomous robotic system designed for the harvesting of cherries. The invention seeks to address the need for increased efficiency and reduced labor costs in cherry picking through automation.
Field of the Invention: The invention applies to the field of agricultural robotics, with a focus on the automation of fruit harvesting processes.
Summary of the Invention: This invention relates to CherryBot, an automated cherry picking robot that employs a sophisticated array of sensors and algorithms to navigate orchard environments, identify ripe cherries, and determine optimal picking strategies. The system is designed to work autonomously, with minimal human intervention.
Brief Description of the Drawings:
• Fig. 1: A side view of CherryBot showcasing the solar panels and manipulator arms.
• Fig. 2: An overhead view of the robot navigating an orchard path.
• Fig. 3: A detailed view of the adaptive gripper engaging a cherry.
• Fig. 4: A diagram of the AI decision-making and control system.
• Fig. 5: An illustrative depiction of the sensor fusion process integrating visual, LiDAR, and infrared data.
Detailed Description: The CherryBot is an innovative robotic system comprising several components that function together for effective cherry harvesting:
1. Autonomous Navigation System: A GPS and sensor-based navigation platform enabling the robot to maneuver through various terrains within cherry orchards.
2. Visual, LiDAR, and Infrared Sensor Suite: This suite includes high-resolution and stereo vision cameras, 3D LiDAR scanners, miniaturized LiDAR for manipulator arm guidance, thermal imaging, and NIR spectroscopy sensors, all integrated through sensor fusion algorithms to accurately identify ripe cherries and their spatial location.
3. Manipulator Arms with Adaptive Grippers: Articulated arms equipped with grippers designed to pick cherries without causing bruising, featuring proximity sensors for collision avoidance.
4. Machine Learning and Image Processing Algorithms: A combination of CNNs, object detection models, and semantic segmentation for cherry detection, alongside path planning algorithms like TSP solvers and dynamic programming for movement efficiency.
5. Control System and AI: A sophisticated AI system that integrates sensor data and controls robotic operations, including a calibration system for maintaining sensor accuracy and a feedback loop for continuous performance improvement.
Claims:
1. A robotic system for harvesting cherries, comprising:
• an autonomous vehicle platform with stabilization features;
• an energy system incorporating solar panels;
• a suite of visual, LiDAR, and infrared sensors for detection and analysis of cherries;
• a plurality of robotic arms with adaptive grippers for picking cherries;
• a machine-learning based detection system trained to identify ripe cherries;
• a navigation system guided by GPS and integrated sensors;
• a collection mechanism for storing harvested cherries.
Additional claims would further specify the configurations and operations of each system component.
Abstract: The CherryBot is a cutting-edge autonomous robot designed to harvest cherries. It uses a combination of visual, LiDAR, and infrared sensors to identify ripe cherries and employs machine learning algorithms for optimal picking strategy determination. This system promises to revolutionize cherry harvesting by improving efficiency and safety while reducing labor costs.
Conclusion, Ramifications, and Scope of Invention: This invention presents a novel solution to cherry harvesting with the potential for adaptation to other fruit-picking applications. Variations in design and method are contemplated, all within the scope of the invention as claimed.
This revised outline includes specific technical details and is crafted to align with patent application requirements. It is essential, however, to consult with a patent attorney to prepare and file an actual application, as the attorney can ensure that the application meets all legal criteria and has the best chance of being granted.
Top of Form
Bottom of Form
Bottom of Form
Prompt
would a law review article on ai and patents be interesting? Please give me a short answer, no more than five sentences
ChatGPT
Yes, a law review article on AI and patents would be quite interesting. It could explore the evolving role of AI in the invention process, the implications for patent law, and the challenges posed by AI-created inventions. Such an article could offer valuable insights into how the legal system might adapt to technological advancements. The intersection of AI and intellectual property is a fertile ground for academic discussion and could influence future legal frameworks. ↑