Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update LLM Options & README Instructions #108

Merged
merged 3 commits into from
Nov 11, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
42 changes: 16 additions & 26 deletions README.md
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
# Autogluon Assistant
# AutoGluon Assistant

[![Python Versions](https://img.shields.io/badge/python-3.8%20%7C%203.9%20%7C%203.10%20%7C%203.11-blue)](https://pypi.org/project/autogluon-assistant/)
[![GitHub license](https://img.shields.io/badge/License-Apache_2.0-blue.svg)](./LICENSE)
Expand All @@ -8,7 +8,7 @@ AutoGluon Assistant (AG-A) provides users a simple interface where they can inpu

## Setup

```
```bash
# create a conda env
conda create -n aga python=3.10
conda activate aga
Expand All @@ -28,7 +28,7 @@ AG-A supports using both AWS Bedrock and OpenAI as LLM model providers. You will
AG-A integrates with AWS Bedrock by default. To use AWS Bedrock, you will need to configure your AWS credentials and region settings, along with the Bedrock-specific API key:

```bash
export BEDROCK_API_KEY="4509..."
export BEDROCK_API_KEY="<your-bedrock-api-key>"
export AWS_DEFAULT_REGION="<your-region>"
export AWS_ACCESS_KEY_ID="<your-access-key>"
export AWS_SECRET_ACCESS_KEY="<your-secret-key>"
Expand All @@ -51,50 +51,38 @@ Important: Free-tier OpenAI accounts may be subject to rate limits, which could

## Usage

We support two ways of using AutoGluoon Assistant: WebUI and CLI.
We support two ways of using AutoGluon Assistant: WebUI and CLI.

### Web UI
The Autogluon Assistant Web UI is a user-friendly application that allows users to leverage the capabilities of the Autogluon-Assistant library through an intuitive web interface.
AutoGluon Assistant Web UI allows users to leverage the capabilities of AG-A through an intuitive web interface.

The web UI enables users to upload datasets, configure Autogluon-Assistant runs with customized settings, preview data, monitor execution progress, view and download results, and supports secure, isolated sessions for concurrent users.
The web UI enables users to upload datasets, configure AG-A runs with customized settings, preview data, monitor execution progress, view and download results, and supports secure, isolated sessions for concurrent users.

#### To run the Autogluon Assistant Web UI:
#### To run the AG-A Web UI:

````
```bash
aga ui

# OR

# Launch Web-UI on specific port e.g. 8888
aga ui --port 8888
```

````

Autogluon Assistant Web UI should now be accessible in your web browser at `http://localhost:8501`

#### Add GPT4 Model to the LLM Option:
If you’d like to add additional GPT4 model to the language model (LLM) dropdown:

1. Navigate to src/autogluon_assistant/WebUI/constants.py
AG-A Web UI should now be accessible in your web browser at `http://localhost:8501` or the specified port.

2. Locate the `LLM_OPTIONS` variable, which looks like this:
````
LLM_OPTIONS = ["Claude 3.5 with Amazon Bedrock"]
````
3. Add "GPT 4o" to the list
````
LLM_OPTIONS = ["Claude 3.5 with Amazon Bedrock", "GPT 4o"]
````

### CLI

Before launching AG-A CLI, prepare your data files in the following structure:

```
└── data # Data files directory
├── train.[ext] # Training dataset (required)
├── test.[ext] # Test dataset (required)
└── description.txt # Dataset and task description (recommended)
```

Note:
- The training and test files can be in any tabular data format (e.g., csv, parquet, xlsx)
- While there are no strict naming requirements, we recommend using clear, descriptive filenames
Expand All @@ -105,7 +93,8 @@ Note:
- Any other relevant information

Now you can launch the AutoGluon Assistant run using the following command:
```

```bash
aga run [NAME_OF_DATA_DIR] --presets [PRESET_QUALITY]
# e.g. aga run ./toy_data --presets best_quality
```
Expand All @@ -119,7 +108,8 @@ You can override specific settings in the YAML configuration defined in the [con
the `config_overrides` parameter with Hydra syntax from the command line.

Here’s an example command with some configuration overrides:
```

```bash
aga run toy_data --config_overrides "feature_transformers=[], autogluon.predictor_fit_kwargs.time_limit=3600"

# OR
Expand Down
2 changes: 1 addition & 1 deletion src/autogluon_assistant/ui/constants.py
Original file line number Diff line number Diff line change
Expand Up @@ -38,7 +38,7 @@
"GPT 4o": "gpt-4o-mini-2024-07-18",
}

LLM_OPTIONS = ["Claude 3.5 with Amazon Bedrock"]
LLM_OPTIONS = ["Claude 3.5 with Amazon Bedrock", "GPT 4o"]

# Provider configuration
PROVIDER_MAPPING = {"Claude 3.5 with Amazon Bedrock": "bedrock", "GPT 4o": "openai"}
Expand Down
4 changes: 2 additions & 2 deletions src/autogluon_assistant/ui/pages/task.py
Original file line number Diff line number Diff line change
Expand Up @@ -113,8 +113,8 @@ def config_time_limit():
@st.fragment
def config_llm():
st.selectbox(
"Choose a LLM model",
placeholder="Choose a LLM model",
"Choose an LLM model",
placeholder="Choose an LLM model",
options=LLM_OPTIONS,
key="_llm",
on_change=store_value,
Expand Down
Loading