• Published on

    Vercel has released the Vercel AI SDK 3.0, making it easier for developers to create user interfaces for language models by using React Server Components.

    The SDK lets developers stream UI components directly from LLMs without the need for heavy client-side JavaScript, making apps more interactive and responsive.

    It also allows developers to give LLMs rich, component-based interfaces so users can better interpret and visualise the outputs from AI models.

    The release is based on the company’s previous development of v0, a generative UI design tool that uses React Server Components to convert text and image prompts to React UIs.

  • Published on

    Amazon Web Services (AWS) has launched Amazon Q, a generative AI tool for answering questions and solving problems using a customer’s own data and systems.

    The tool can be tailored to individual users, with more than 40 connectors to popular data sources, including Salesforce, Dropbox, ServiceNow and Zendesk.

    Amazon Q will be available through the AWS Management Console, and has been specifically developed to be accessible to companies of all sizes.

    The product will be rolled out in two flavours: Amazon Q in Connect, an application for customer service workers that will suggest answers to customer service workers, and Amazon Q in AWS Supply Chain, which will offer supply chain workers answers to complex questions based on their data.

  • Published on

    GitHub has announced a number of updates at its GitHub Universe 2023 event, including the general availability of GitHub Copilot Chat.

    The tool enables developers to write and understand code using natural language, and will be available as part of an existing GitHub Copilot subscription from December 2023.

    As well as integrating the tool into its web and mobile apps, GitHub has also launched GitHub Copilot Enterprise, which allows developers to quickly get up to speed with a codebase, search through and build documentation, and review pull requests.

    It costs $39 per user per month and will be generally available in February 2024.

    GitHub has also announced its GitHub Copilot Partner Program, which will see the integration of the tool with third-party developer tools and services, and new AI-powered security features for GitHub Advanced Security, including code scanning autofix and secret scanning.

  • Published on

    AI models are becoming more prevalent and developers require appropriate tools to effectively utilise them.

    Current approaches to working with LLMs, such as prompting and fine-tuning, are insufficient, since developers do not sufficiently understand how the models produce outputs from their inputs.

    To address this, Martian has developed a model mapping technique to turn transformers into programs, allowing developers to understand how models work and make use of them more effectively.

    The first application of this technique is the model router, which can determine the best LLM to use for each query and route it in real time to achieve the best performance at the lowest cost.

    This is the first commercial application of large-scale AI interpretability and achieves better results than GPT-4 at a lower cost.

  • Published on

    GPT Researcher is a tool that uses AI to undertake online research. It aims to address the problems of speed, determinism and reliability that researchers often face.

    The tool uses “planner” and “execution” agents to generate research questions, which are then used to search for relevant information online. This information is then filtered and aggregated into a research report.

    The tool is intended to produce factual and unbiased research whilst offering customisation options and using over 20 web sources to reduce the risk of incorrect information.

    The average research task takes three minutes and costs around $0.1.

  • Published on

    PrivateGPT is an API that allows users to ask questions about their documents using the power of large language models (LLMs) without risking a data leak.

    The project is production-ready and provides two logical blocks: a high-level API that manages ingestion, splitting, metadata extraction, embedding generation and storage, and a low-level API for implementing more complex pipelines.

    It also offers a Gradio UI client to test the API and tools such as a bulk model download script and a documents folder watch.

    Based on the LlamaIndex framework, PrivateGPT is designed to be easily extended and adapted to suit user needs, with dependency injection, usage of abstractions, and simplicity of design.