Skip to content

Commit

Permalink
feat: Organize Inference Files - Part 1 (#216)
Browse files Browse the repository at this point in the history
This is the first part towards the new inference pipeline - organizing
inference files into their own folder.

This way we can easier trigger build pipelines (edits to
presets/models/inference) as well as maintain/add new types of inference
  • Loading branch information
ishaansehgal99 authored Jan 26, 2024
1 parent d334f26 commit ecc56d1
Show file tree
Hide file tree
Showing 11 changed files with 208 additions and 171 deletions.
1 change: 0 additions & 1 deletion pkg/inference/preset-inference-types.go
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,6 @@ var (
}

DefaultAccelerateParams = map[string]string{
"config_file": DefaultConfigFile,
"num_processes": DefaultNumProcesses,
"num_machines": DefaultNumMachines,
"machine_rank": DefaultMachineRank,
Expand Down
27 changes: 0 additions & 27 deletions presets/models/falcon/config.yaml

This file was deleted.

118 changes: 0 additions & 118 deletions presets/models/falcon/inference-api.py

This file was deleted.

Original file line number Diff line number Diff line change
@@ -1,21 +1,21 @@
# Copyright (c) Microsoft Corporation.
# Licensed under the MIT license.
from fastapi import FastAPI, HTTPException
import uvicorn
from pydantic import BaseModel
from typing import Optional
import argparse
import functools
import multiprocessing
import multiprocessing.pool
import os
import signal
import sys
import threading
import functools
from typing import Optional

from llama import Llama
import torch
import sys
import signal
import os
import torch.distributed as dist
import argparse
import uvicorn
from fastapi import FastAPI, HTTPException
from llama import Llama
from pydantic import BaseModel

# Setup argparse
parser = argparse.ArgumentParser(description="Llama API server.")
Expand Down
Original file line number Diff line number Diff line change
@@ -1,21 +1,21 @@
# Copyright (c) Microsoft Corporation.
# Licensed under the MIT license.
from fastapi import FastAPI, HTTPException
import uvicorn
from pydantic import BaseModel
from typing import Optional
import argparse
import functools
import multiprocessing
import multiprocessing.pool
import os
import signal
import sys
import threading
import functools
from typing import Optional

from llama import Llama
import torch
import sys
import signal
import os
import torch.distributed as dist
import argparse
import uvicorn
from fastapi import FastAPI, HTTPException
from llama import Llama
from pydantic import BaseModel

# Setup argparse
parser = argparse.ArgumentParser(description="Llama API server.")
Expand Down
Loading

0 comments on commit ecc56d1

Please sign in to comment.