Modern, open-source Jupyter alternative.
Try it here »
Discord
·
Website
·
Issues
·
Contact
Pretzel.AI.Overview.with.Subtitles.mp4
Pretzel is a fork of Jupyter with the goal to improve Jupyter's capabilities. We've added AI code generation and editing, inline tab completion, sidebar chat and error fixing to Jupyter for now with a lot more to come.
Switching to Pretzel from Jupyter is extremely easy since it's simply an improved version of Jupyter. All of your Jupyter config, settings, keybindings, and extensions will work out of the box.
- Installation:
pip install pretzelai
then runpretzel lab
to open the web interface. OR, use our free hosted version: pretzelai.app- See this if you get errors during installation
- Simply start typing in a cell to get inline tab completions
- In any Jupyter cell, click “Ask AI” or press Cmd+K (Mac) / Ctrl+K (Linux/Windows) to prompt AI
- Use the AI Sidebar with Ctrl+Cmd+B (Mac) or Ctrl+Alt+B (Linux/Windows) to chat with AI, generate code, and ask questions.
- Type
@
to trigger auto-complete with function and variable names. - To use your own model (OpenAI, Anthropic/Claude, Ollama or Groq), see the Configuration section
Our roadmap includes building features such as:
- Native AI code generation and understanding features similar to Cursor
- Frictionless realtime collaboration: pair-programming, comments, version history, etc.
- SQL support (both in code cells and as a standalone SQL IDE)
- Visual analysis builder (see more here)
- VSCode like code-writing experience using Monaco
- 1-click dashboard creation and sharing from Jupyter notebooks
Install Pretzel with pip
pip install pretzelai
For conda, install pip first with conda install pip
and then pip install pretzelai
.
Then, start Pretzel with:
pretzel lab
You'll be able to access the Pretzel interface via the provided URL.
To use your own AI model, see the Configuration section.
If you're having trouble installing Pretzel (for eg on Windows), you can run it in a Docker container.
- Create a Dockerfile:
FROM python:3.9-slim
RUN apt-get update && apt-get install -y \
build-essential \
gcc \
&& rm -rf /var/lib/apt/lists/*
WORKDIR /root/pretzel
RUN pip install pretzelai
EXPOSE 8888
CMD ["pretzel", "lab", "--ip=0.0.0.0", "--allow-root", "--notebook-dir=/root/pretzel", "--ServerApp.allow_remote_access=True", "--ServerApp.token=''", "--no-browser"]
-
In the same folder where you have your Dockerfile, run
docker build -t pretzel .
-
To run pretzel, you can run:
docker run --name pretzel -p 8888:8888 pretzel
and once the container is running, you can access it athttp://localhost:8888/lab
. To stop the container, pressCtrl + C
followed bydocker stop pretzel
.
If you want to access your local folder in Pretzel, you can run: docker run --rm -p 8888:8888 -v $(pwd):/root/pretzel pretzel
- this will map your current directory to the docker container's /root/pretzel
folder. Make sure Docker has access to your current directory.
To update Pretzel to the latest version, just rebuild the Docker image with the --no-cache flag: docker build --no-cache -t pretzel .
and now you can run docker like step 3.
You can use this Dockerfile to build and run the bleeding edge version. Follow the steps (starting at step 2) in the section Running within a docker container
If you get an error during installation that looks like this:
Failed to build installable wheels for some pyproject.toml based projects (pystemmer)
That means the installation failed to install the dependency PyStemmer
. This usually happens because you don't have the right build tools installed. To fix this:
- On Windows: Install Miscrosoft build tools from here. Click on the Download Build Tools button and then install it.
- On Ubuntu (and Debian flavoured systems): Install the required build tools by running the following command in your terminal:
sudo apt-get update && sudo apt-get install build-essential python3-dev
- On macOS:
brew install gcc
. If this doesn't work, you may also need to runxcode-select --install
Once this is done, you should be able to pip install pretzelai
to install Pretzel.
- Start typing in a cell to get inline tab completions
- Wait for 1 second to trigger completions. You'll see a little spinner just before the blue Ask AI button in the cell
The default Pretzel AI Server uses Mistral's Codestral but you can switch the inline completion model in Pretzel AI Settings. See the configuration section.
- In a cell, press
Cmd+K
(Mac) /Ctrl+K
(Windows/Linux) or click "Ask AI" to open AI prompt textbox and write your code generation/editing instruction- Type
@
to get a dropdown of available variables in your session. Adding this@vairable
to the prompt will send its value to the AI - We automatically send relevant code in the current notebook as context to the AI
- Type
- If there's existing code in a cell, the prompt will edit the existing code
- If you select/highlight some code in the cell, only the selected code will be edited
- You can accept/reject the response or edit your prompt if you want to re-submit with modifications
- Use ↑ / ↓ to cycle through prompt history
- Use
Ctrl+Cmd+B
(Mac) /Ctrl+Alt+B
(Linux/Windows) or the Pretzel Icon on the right sidebar to activate the AI Sidebar - You can ask questions, generate code, or search for existing code
- The AI always uses the code in the active cell as context. If you highlight some code in the active cell, only the highlighted code will be used as context
- We automatically send relevant code in the current notebook as context to the AI
- You can also use the
@
syntax to refer to variables and dataframes in memory, similar to the notebook cells
Example uses of AI Sidebar:
- "Modify the function
my_function
to be more efficient" ← this will find the functionmy_function
in the whole notebook and modify it - "Where is the code that removes outliers"? ← this will search the notebook for code that removes outliers and show it to you
- (While cursor is in a cell with some code) "Can you explain what this code does?" ← this will explain the code in the current cell
- "Plot a histogram of @df for age" ← this will generate code to plot a histogram of the 'age' column from the dataframe named 'df'
- "Calculate the mean revenue in @salesdata for each product type" ← _this will generate code to calculate mean revenue by product type using the
sales_data
dataframe
- Put your cursor either on an empty line or an existing line of code. Bring up the AI prompting text box with Cmd+K
- Start your prompt with the word
inject
orij
(case-insensitive) - this tells the AI to only add new code and not edit the existing code in the cell - Code will be added one line below where your cursor was placed
- When there's an error, you'll see a button on top-right "Fix Error with AI". Click it try fixing the error
Pretzel works out-of-the-box, no configuration needed. We DO NOT store any code or data you send to the Pretzel AI Server.
Still, if you want to use a different AI model, you can configure Pretzel to use AI models from multiple different vendors, including local models. In this case, the AI request goes directly from your computer to the AI service (OpenAI, Anthropic etc).
- Open the
Settings
menu in the top menubar, then clickPretzel AI Settings
- Under AI Settings you can pick which AI models to use for chat and for inline completion (see screenshot below)
- We recommend using GPT-4 class models (e.g., GPT-4 Turbo, GPT-4o, Claude-3.5 Sonnet, Claude-3 Opus and Llama-3.1 405B) for the best performance
Under the section Configure AI Services, you can
- Enable or disable certain AI services (for eg, Azure and Ollama are disabled by default but can be enabled)
- Enter your API key or URL as needed for each service
Remember to save your settings after making changes. Pretzel will validate your configuration to ensure everything is set up correctly.
Please note: We haven't tested Azure Enterprise OpenAI models yet. If you find any bugs, please report them in GitHub issues and we'll fix them ASAP.
- Please report bugs here: https://github.com/pretzelai/pretzelai/issues
- Have any feedback? Any complains? We'd love feedback: [email protected]
The original Jupyter documentation is available here and the Jupyterlab README is available here.
We collect no personal information. We use basic telemetry for only the AI features we've built - for example, when you click on "Ask AI", we receive an event that someone clicked on "Ask AI". We only associate an anonymous ID to your user. If you allow cookies, that helps us tell that it's the same user across multiple browser sessions (which is very helpful!). If you don't allow cookies, every time you open a browser, you're a new anonymous user to us.
We also collect prompts (but not the responses) for the AI features we've built. This can be turned off in the settings (Settings > Pretzel AI > Uncheck Prompt Telemetry) but we'd really appreciate if you didn't - this is very helpful in improving our prompts.
We do not collect any code whatsoever. Even when you use Pretzel's cloud AI server for completions, we don't store any of this code.
If you use the hosted version of Pretzel (https://pretzelai.app), we create a user for you based on your email address. You can always simply log-in and delete any data you may have stored on our hosted server. We make no backups or copies of your data.
Our hosted server is free to use. However, we will delete your data and your account 30 days after your last login. If you'd like to delete your account sooner, please email us at [email protected] with the subject line "Account Deletion" and we'll delete your account immediately.
Q. What happened to the old version of Pretzel AI - the visual, in-browser data manipulation tool?
A. It's available in the pretzelai_visual
folder here. Please see this PR for more info.
Q. What AI model does Pretzel use?
A. Pretzel uses different AI models for various tasks:
-
Default model: GPT-4o
- Offers a good balance between speed and quality
- Can be changed in Pretzel Settings if you're using your own API key
-
Inline completions: Mistral's Codestral model
- Excellent for code completion
- Very fast performance (22B parameter model)
-
Fallback option:
- If you're using your own API key without providing a Mistral API Key, Pretzel will use GPT-4o for inline completions as well
We're continuing to experiment with models and supporting local models and Anthropic's Claude is at the top of our list.
Q. What about feature X?
A. There's a ton we want to build. Please open an issue and tell us what you want us to build!
Q. Where's the roadmap?
A. We have a rough roadmap at the top of this README. There are many features we'd like to build, but there's just two of us. So, we're collecting feedback about what would be most helpful. Please open an issue or just email us with your feedback! Based on what we find, we'll prioritize our roadmap.
Q. Why are you using the AGPL license? Or, why not use MIT/BSD3 licenses?
A. Our goal with building Pretzel is to make an amazing data tool that is free for both individuals and companies to use. That said, we are a two-person startup - and we don't want some third party to just take our code and sell a hosted version of it without giving back to the community. Jupyter code is licensed as BSD-3 and if we keep our new code BSD-3 licensed, there would be no way to stop a third party from doing this. As a result, we went with the AGPLv3 license for all the new code. This ensures that if someone else does want to take our code and sell it (SaaS or otherwise), they have to open-source all of their modifications under AGPLv3 as well.
Q. Why a fork of Jupyter? Why not contribute into Jupyter directly?
A. This deserves a longer answer but here's the short answer: We've set out to make the new de-facto, modern, open-source data tool. Initially, we wanted to start from scratch. However, after talking to several data professionals, we realized it will be very hard to get people to switch to a new tool, no matter how good. The best way to get people to switch is to not have them switch at all. That's why we decided to fork Jupyter - for the near zero switching costs. Also, Jupyter is a mature product, and we're shipping feature really fast - frankly, at the pace we're shipping features, the code we write won't be accepted into the Jupyter codebase 😅. There are also many downsides to this decision - we've had to spend considerable time understanding the whole Jupyter ecosystem and multiple codebases, the complex release processes, the various APIs etc. However, we think this is the right decision for us.
Q. My company is worried about using an AGPLv3 licensed tool. What can I do?
A. The AGPL is a barrier ONLY IF you're modifying Pretzel AND redistributing it to the public. If you're simply using it as a tool in your company (even with modifications), the AGPL DOES NOT ask you to share your code. Still, if AGPL is an issue for you, please contact us, and we can figure out something that works.
Q. How are you planning on making money? OR, how are you free? I'm worried that you'll make this tool paid in the future.
A. We're planning on selling a hosted version of the tool to companies to make money. This hosted version will probably have some company specific features that individuals don't want or need such as data access controls, connectors for data sources, integration with GitHub, hosted and shareable dashboard, scalable and on-demand compute for large data jobs etc. We will not retroactively make Pretzel's individual version paid.