JAN: This 100% LOCAL AI ASSISTANT is better than ChatGPT (works w/ RAG, Local Models, Groq & OpenAI)

AICodeKing
10 May 202409:26

TLDRJan Jan is a versatile desktop application that allows users to run open-source AI models locally for enhanced privacy, or connect to online models via API keys from platforms like OpenAI or Gro for broader capabilities. It's open-source, user-friendly, and supports cross-platform use on Mac, Linux, and Windows. Jan offers the flexibility to switch between local and remote models, stores conversations offline, and provides API endpoints for custom applications. It also supports extensions for additional functionality and has built-in engines for inference. The app is easy to install and operate, with a sleek interface that guides users through model selection, chat interactions, and advanced settings. It's an all-in-one solution for those looking to leverage AI without the complexities of terminal commands or multiple configurations.

Takeaways

  • 🌐 Jan is a desktop app that allows running open-source AI models locally and also supports connecting to online models like OpenAI or Groq.
  • πŸ”’ Jan emphasizes enhanced privacy by enabling local model execution, which is beneficial for sensitive conversations.
  • πŸ”Œ If no internet connection is available or not desired, Jan's local model feature ensures private and secure conversations.
  • πŸ”€ Flexibility is provided by allowing users to switch between local and remote models according to their needs.
  • πŸ“‚ All conversations with Jan are stored offline, and it is cross-platform, available on Mac, Linux, and Windows.
  • πŸ“‘ Jan exposes API endpoints that are compatible with OpenAI, allowing integration with custom applications.
  • 🧩 It offers an extension option for adding custom plugins or integrating with other tools and services.
  • πŸ“š Jan can work with various text files, including PDFs and documents, and has two built-in engines for inference: llama CPP and tensor RT.
  • πŸš€ Jan's dual engine approach provides flexibility and options for model inference.
  • πŸ”§ Users can connect Jan with LM Studio or other endpoints if needed, and it's easy to install with a one-click process.
  • πŸ”‘ Jan allows setting up API keys for services like Gro, which is useful for accessing additional capabilities without advanced hardware.

Q & A

  • What is Jan Jan?

    -Jan Jan is a desktop app that allows users to run open-source AI models locally and also connect to online models like OpenAI or Groq using API keys, providing a unified platform to interact with various AI models.

  • How does Jan Jan enhance privacy for users?

    -Jan Jan enhances privacy by enabling users to run AI models like Llama or Mistral directly on their device, eliminating the need for an internet connection. This is particularly useful for sensitive or confidential conversations.

  • What are the benefits of using local models in Jan Jan?

    -Using local models in Jan Jan ensures that conversations remain private and secure, as no data is transmitted over the internet. It also allows for the use of AI capabilities without the need for advanced hardware.

  • Can Jan Jan be used across different operating systems?

    -Yes, Jan Jan is cross-platform and is available on Mac, Linux, and Windows, making it accessible to a wide range of users.

  • How does Jan Jan's API endpoint feature work?

    -Jan Jan exposes API endpoints that can be used for custom applications or other AI applications. These endpoints are open AI compatible, allowing integration with any system that supports open AI models.

  • What is the dual engine approach in Jan Jan?

    -The dual engine approach in Jan Jan refers to the use of two built-in engines for inference: llama CPP and tensor RT, LM. This provides users with more flexibility and options when it comes to model inference.

  • How can users install and use a local model in Jan Jan?

    -Users can install a local model in Jan Jan by visiting the app's website, downloading the installation file for their operating system, and following the setup instructions. They can then explore the Hub to find and download models, or paste a Hugging Face link to automatically download a model.

  • What is the purpose of the 'New Thread' option in Jan Jan?

    -The 'New Thread' option in Jan Jan allows users to create new conversation threads, which can be useful for organizing different discussions or inquiries within the app.

  • How can users integrate Jan Jan with their custom tools and services?

    -Users can integrate Jan Jan with their custom tools and services by using the exposed API endpoints or by adding custom plugins through the extensions option.

  • What is the significance of the 'Retrieval' feature in Jan Jan?

    -The 'Retrieval' feature, which is part of the RAG (Retrieval-Augmented Generation) capabilities, allows users to attach a file and ask questions about its content. This is particularly useful for extracting information from documents or PDFs.

  • How does Jan Jan handle conversations with online models?

    -Jan Jan handles conversations with online models by allowing users to connect their OpenAI or Groq API keys. Once connected, users can select online models like Gro Llama 3 from the models dropdown and engage in fast-paced conversations.

  • What are the advantages of using Jan Jan over other AI model platforms?

    -Jan Jan offers a one-click installation process, the ability to use local and online models, cross-platform accessibility, API endpoint exposure for custom integrations, and the flexibility to switch between different models and providers easily.

Outlines

00:00

πŸ–₯️ Jan App Overview and Local Model Usage

The video introduces Jan, a desktop application that facilitates the local running of open-source AI models and the connection of online models via API keys. Jan is open-source, user-friendly, and supports a range of functionalities including local model execution for privacy, cross-platform availability, API endpoint exposure, and extension options. The video demonstrates how to install Jan, select a local model like Llama 3, and interact with it through a chat interface. It also covers how to customize the model's settings and create new threads for different conversations.

05:00

πŸ”Œ Advanced Settings and Online Model Integration

This paragraph delves into the advanced settings of Jan, including the ability to upload custom models, enabling experimental features, and toggling GPU acceleration. It also explains how to set up API endpoints for models and manage HTTPS proxies. The video guides viewers on how to configure API keys for online models like Gro and demonstrates the process of sending messages using these online models. Additionally, it introduces the Retrieval-Augmented Generation (RAG) feature, which allows users to attach documents and ask questions about them. The video concludes with a call to action for viewers to share their thoughts and support the channel.

Mindmap

Keywords

πŸ’‘Jan

Jan is a desktop application that serves as a local AI assistant, allowing users to run open-source models directly on their devices. It is designed to be user-friendly, requiring no coding or terminal usage. In the context of the video, Jan is highlighted for its ability to connect to various AI models, ensuring privacy for sensitive conversations by running them locally, or leveraging the power of online models when needed.

πŸ’‘Local Models

Local models refer to AI models that run directly on the user's device without the need for an internet connection. This is significant for privacy and security, as it allows users to have confidential conversations without the risk of data being intercepted over the network. In the video, Jan's support for local models is emphasized as a key feature, particularly useful for handling sensitive information.

πŸ’‘Open Source

Open-source software is a type of software where the source code is made available to the public, allowing anyone to view, use, modify, and distribute the software. Jan is described as fully open source, which means its code is accessible, and users can contribute to its development or modify it to suit their needs. This is important as it fosters a community-driven approach to improving the application.

πŸ’‘API Keys

API keys are unique identifiers used to authenticate users when accessing an API (Application Programming Interface). In the video, it is mentioned that users can connect their OpenAI or Gro API keys to Jan, which allows them to interact with online models. This is crucial for users who want to utilize the capabilities of these online models without having to host the models themselves.

πŸ’‘Crossplatform

Crossplatform refers to the ability of an application to function on multiple operating systems without the need for specific versions for each platform. Jan is described as being crossplatform, available on Mac, Linux, and Windows, which makes it accessible to a wider range of users regardless of their preferred operating system.

πŸ’‘API Endpoints

API endpoints are specific URLs that are used to access different services or data from an API. In the context of the video, Jan exposes API endpoints that can be utilized for custom applications or other AI applications. This feature is important as it allows for integration with other tools and services, enhancing the functionality of Jan.

πŸ’‘Extensions

Extensions in the context of Jan refer to additional features or functionalities that can be added to the core application to enhance its capabilities. The video mentions that users can set up extensions to add custom plugins or integrate Jan with other tools and services. This is beneficial as it allows for a more personalized and flexible user experience.

πŸ’‘Inference Engines

Inference engines are components of AI applications that perform the task of drawing conclusions or making decisions based on input data. Jan has two built-in inference engines: llama CPP and tensor RT, LM. The dual engine approach is beneficial as it provides users with more flexibility and options when it comes to model inference, allowing them to choose the most suitable engine for their needs.

πŸ’‘Llama 3 Model

The Llama 3 Model is an AI model that can be run locally on a user's device through the Jan application. It is mentioned in the video as an example of a local model that can be downloaded and used within Jan. The use of the Llama 3 Model illustrates the application's ability to handle sophisticated AI models for various tasks, such as generating responses to user inputs.

πŸ’‘Retrieval-Augmented Generation (RAG)

Retrieval-Augmented Generation (RAG) is a feature that allows AI models to retrieve relevant information from a database or a document before generating a response. In the video, it is shown how to enable RAG features in Jan, which enhances the model's ability to provide informed and contextually relevant answers by incorporating data from attached files.

πŸ’‘Gro API

The Gro API is a service that provides access to AI models for developers. In the video, it is demonstrated how to set up the Gro API key within Jan, which allows users to connect to and utilize the capabilities of the Gro models. The mention of the Gro API highlights the application's compatibility with various online AI model providers, offering users more options for model usage.

Highlights

JAN is a desktop app that allows running open-source AI models locally and connecting with online models via API keys.

It is fully open source, easy to install, and does not require terminal, coding, or configuration.

JAN supports running local AI models like Llama or Mistral for enhanced privacy without an internet connection.

The app offers the ability to switch between local and remote models like OpenAI or Gro for flexibility.

All conversations with JAN are stored offline and the app is cross-platform, available on Mac, Linux, and Windows.

JAN exposes API endpoints that are compatible with OpenAI models for custom applications.

The app has an extensions option for adding custom plugins or integrating with other tools and services.

JAN uses two built-in engines for inference: llama CPP and tensor RT LM.

Users can connect JAN with LM Studio or other endpoints if needed.

The interface allows for easy installation of local models and provides options to explore and download models.

JAN enables users to rename threads, change instructions, and select models for conversation.

Advanced settings for models can be adjusted, including temperature and max tokens.

The app provides options to create new threads and install models from the hubs.

JAN allows for setting up API endpoints for the model and checking logs.

Settings include options for managing installed models, experimental features, GPU acceleration, and API configurations.

Users can set up and use JAN with various LLM providers like Gemini, OpenAI, Gro, or local models.

JAN supports RAG (Retrieval-Augmented Generation) features for attaching files and asking questions about the content.

The app is designed to be user-friendly, eliminating the need for complex terminal configurations.

JAN is a one-click install solution that centralizes the management of multiple LLM providers.