Document supported models (#127)
This commit is contained in:
40
docs/source/models/supported_models.rst
Normal file
40
docs/source/models/supported_models.rst
Normal file
@@ -0,0 +1,40 @@
|
||||
.. _supported_models:
|
||||
|
||||
Supported Models
|
||||
================
|
||||
|
||||
CacheFlow supports a variety of generative Transformer models in `HuggingFace Transformers <https://github.com/huggingface/transformers>`_.
|
||||
The following is the list of model architectures that are currently supported by CacheFlow.
|
||||
Alongside each architecture, we include some popular models that use it.
|
||||
|
||||
.. list-table::
|
||||
:widths: 25 75
|
||||
:header-rows: 1
|
||||
|
||||
* - Architecture
|
||||
- Models
|
||||
* - :code:`GPT2LMHeadModel`
|
||||
- GPT-2
|
||||
* - :code:`GPTNeoXForCausalLM`
|
||||
- GPT-NeoX, Pythia, OpenAssistant, Dolly V2, StableLM
|
||||
* - :code:`LlamaForCausalLM`
|
||||
- LLaMA, Vicuna, Alpaca, Koala
|
||||
* - :code:`OPTForCausalLM`
|
||||
- OPT, OPT-IML
|
||||
|
||||
If your model uses one of the above model architectures, you can seamlessly run your model with CacheFlow.
|
||||
Otherwise, please refer to :ref:`Adding a New Model <adding_a_new_model>` for instructions on how to implement support for your model.
|
||||
Alternatively, you can raise an issue on our `GitHub <https://github.com/WoosukKwon/cacheflow/issues>`_ project.
|
||||
|
||||
.. tip::
|
||||
The easiest way to check if your model is supported is to run the program below:
|
||||
|
||||
.. code-block:: python
|
||||
|
||||
from cacheflow import LLM
|
||||
|
||||
llm = LLM(model=...) # Name or path of your model
|
||||
output = llm.generate("Hello, my name is")
|
||||
print(output)
|
||||
|
||||
If CacheFlow successfully generates text, it indicates that your model is supported.
|
||||
Reference in New Issue
Block a user