Voice is the new computer interface for mobile and ubiquitous computing.
Virtual assistants are like browsers to the voice web: they are fast becoming primary gateways between the internet and our private lives.
A natural-language programming interface for virtual assistants lets end users automate personal long-tail tasks easily.
Stanford Open Virtual Assistant Lab (OVAL) is dedicated to advancing and democratizing virtual assistant technology, while protecting public interest in privacy, open knowledge access, and open competition.
OVAL is supported by the National Science Foundation under Grant No. 1900638, and by the Alfred P. Sloan Foundation under Grant No. G-2020-13938.
We have launched an initiative in July to create an open-source virtual assistant infrastructure to support experimentation in research and to provide a basis of collaboration in the industry. This is made possible by a grant from Alfred P. Sloan Foundation with the goals of protecting open access to knowledge and to protect privacy.
We will be making beta releases throughout this year, with the goals of delivering in one year:
We are also looking for a couple of talented engineers to join the team. Job announcement
Our current partners include:
A paradigm shift is necessary to advance and democratize virtual assistants. Existing assistants built with Virtual Assistant 1.0 technology has limited extensibility and is powered by multibillion-dollar manual annotation efforts. The Virtual Assistant 2.0 technology we developed is grounded in formal programming language semantics and is powered by open-source toolsets and knowledge bases intended to help millions of developers create neural-based conversational agents for their applications cost-effectively. This technology has been shown to outperform commercial systems on long-tail questions. Please follow our technology development here.
We are actively pushing the state of the art on conversational agents. Topics include: improvement in accuracy, scope, and scalability; recovering from errors; low-cost localization; social conversations; better development tools; multimodal assistants. Check this out here.
Our virtual assistant protects privacy by letting users keep private data on their own devices. The code is separated from data. All the code for the skills are stored on the nonproprietary Thingpedia skill repository, which is open to all assistants and contains no personal information. While natural language commands are translated in the cloud, the translated commands are executed locally on the client side to provide privacy. Thus, all the personal account credentials and data are only stored on the users’ devices, and not accessible by any other third party. Local natural language processing is possible once our neural network is trained. Similarly, all the data from the home IoT devices never leave the house.
Our assistant architecture also helps users share data with privacy. It has a federated architecture, just like email; users can use different virtual assistants, which interoperate in sharing data using a standard secure communication protocol.
Our approach is to continuously release working prototypes, under the name of Almond. The current stable version of Almond is available at almond.stanford.edu. All released software and datasets can be downloaded from the Releases page.
Almond 2.0, to be released in the fall, is the first prototype that uses our new dialogue technology. In preparation for the 2.0 release, we are releasing Almond 1.99, which has been tested mainly on the Spotify skill for premium accounts. We are actively improving the model by adding better entity recognition. Check out the Release planning page for more details of the schedule.
See here for the paper abstracts.
We thank them for their valuable contribution.