You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+25Lines changed: 25 additions & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -87,6 +87,31 @@ Visit our [documentation](https://vllm.readthedocs.io/en/latest/) to learn more.
87
87
We welcome and value any contributions and collaborations.
88
88
Please check out [CONTRIBUTING.md](./CONTRIBUTING.md) for how to get involved.
89
89
90
+
## Sponsors
91
+
92
+
vLLM is a community project. Our compute resources for development and testing are supported by the following organizations. Thank you for your support!
93
+
94
+
<!-- Note: Please sort them in alphabetical order. -->
95
+
<!-- Note: Please keep these consistent with docs/source/community/sponsors.md -->
96
+
97
+
- a16z
98
+
- AMD
99
+
- Anyscale
100
+
- AWS
101
+
- Crusoe Cloud
102
+
- Databricks
103
+
- DeepInfra
104
+
- Lambda Lab
105
+
- NVIDIA
106
+
- Replicate
107
+
- Roblox
108
+
- RunPod
109
+
- Trainy
110
+
- UC Berkeley
111
+
- UC San Diego
112
+
113
+
We also have an official fundraising venue through [OpenCollective](https://opencollective.com/vllm). We plan to use the fund to support the development, maintenance, and adoption of vLLM.
114
+
90
115
## Citation
91
116
92
117
If you use vLLM for your research, please cite our [paper](https://arxiv.org/abs/2309.06180):
Copy file name to clipboardExpand all lines: csrc/attention/attention_generic.cuh
+10-9Lines changed: 10 additions & 9 deletions
Original file line number
Diff line number
Diff line change
@@ -1,5 +1,6 @@
1
1
/*
2
-
* Adapted from https://github.com/NVIDIA/FasterTransformer/blob/release/v5.3_tag/src/fastertransformer/kernels/decoder_masked_multihead_attention_utils.h
0 commit comments