text stringlengths 55 456k | metadata dict |
|---|---|
[b](b)
[b.md](b.md)
[./b.md](./b.md)
[/b.md](/b.md) `[/b.md](/b.md)`
[b#header1](b#header1)
```
[b](b)
```
~~~
[b](b)
~~~
// Indented code
[b](b) | {
"source": "voideditor/void",
"title": "extensions/markdown-language-features/test-workspace/a.md",
"url": "https://github.com/voideditor/void/blob/main/extensions/markdown-language-features/test-workspace/a.md",
"date": "2024-09-11T02:37:00",
"stars": 10433,
"description": null,
"file_size": 160
} |
The file `TypeScript.tmLanguage.json` and `TypeScriptReact.tmLanguage.json` are derived from [TypeScript.tmLanguage](https://github.com/microsoft/TypeScript-TmLanguage/blob/master/TypeScript.tmLanguage) and [TypeScriptReact.tmLanguage](https://github.com/microsoft/TypeScript-TmLanguage/blob/master/TypeScriptReact.tmLan... | {
"source": "voideditor/void",
"title": "extensions/typescript-basics/syntaxes/Readme.md",
"url": "https://github.com/voideditor/void/blob/main/extensions/typescript-basics/syntaxes/Readme.md",
"date": "2024-09-11T02:37:00",
"stars": 10433,
"description": null,
"file_size": 1350
} |
# vscode-wasm-typescript
Language server host for typescript using vscode's sync-api in the browser.
## Getting up and running
To test this out, you'll need three shells:
1. `npm i` for vscode itself
2. `npm run watch-web` for the web side
3. `node <root>/scripts/code-web.js --coi`
The last command will open a bro... | {
"source": "voideditor/void",
"title": "extensions/typescript-language-features/web/README.md",
"url": "https://github.com/voideditor/void/blob/main/extensions/typescript-language-features/web/README.md",
"date": "2024-09-11T02:37:00",
"stars": 10433,
"description": null,
"file_size": 2064
} |
# Integration test
## Compile
Make sure to run the following commands to compile and install dependencies:
cd test/integration/browser
npm i
npm run compile
## Run (inside Electron)
scripts/test-integration.[sh|bat]
All integration tests run in an Electron instance. You can specify to run the test... | {
"source": "voideditor/void",
"title": "test/integration/browser/README.md",
"url": "https://github.com/voideditor/void/blob/main/test/integration/browser/README.md",
"date": "2024-09-11T02:37:00",
"stars": 10433,
"description": null,
"file_size": 1111
} |
# First
# Second
[b](/b.md)
[b](../b.md)
[b](./../b.md) | {
"source": "voideditor/void",
"title": "extensions/markdown-language-features/test-workspace/sub/c.md",
"url": "https://github.com/voideditor/void/blob/main/extensions/markdown-language-features/test-workspace/sub/c.md",
"date": "2024-09-11T02:37:00",
"stars": 10433,
"description": null,
"file_size": 56
... |
<!-- Should highlight math blocks -->
$$
\theta
$$
**md**
$$
\theta{ % comment
$$
**md**
$$
\relax{x}{1} = \int_{-\infty}^\infty
\hat\xi\,e^{2 \pi i \xi x}
\,d\xi % comment
$$
**md**
$
x = 1.1 \int_{a}
$
**md**
$
\begin{smallmatrix}
1 & 2 \\
4 & 3
\end{smallmatrix}
$
$
x = a_0 + \frac{1}{a_1 + \... | {
"source": "voideditor/void",
"title": "extensions/vscode-colorize-tests/test/colorize-fixtures/md-math.md",
"url": "https://github.com/voideditor/void/blob/main/extensions/vscode-colorize-tests/test/colorize-fixtures/md-math.md",
"date": "2024-09-11T02:37:00",
"stars": 10433,
"description": null,
"file_... |
# h
<pre><code>
# a
</code></pre>
# h
<pre>
# a
a</pre>
# h | {
"source": "voideditor/void",
"title": "extensions/vscode-colorize-tests/test/colorize-fixtures/test-33886.md",
"url": "https://github.com/voideditor/void/blob/main/extensions/vscode-colorize-tests/test/colorize-fixtures/test-33886.md",
"date": "2024-09-11T02:37:00",
"stars": 10433,
"description": null,
... |
# Header 1 #
## Header 2 ##
### Header 3 ### (Hashes on right are optional)
## Markdown plus h2 with a custom ID ## {#id-goes-here}
[Link back to H2](#id-goes-here)
### Alternate heading styles:
Alternate Header 1
==================
Alternate Header 2
------------------
<!-- html madness -->
<div class=... | {
"source": "voideditor/void",
"title": "extensions/vscode-colorize-tests/test/colorize-fixtures/test.md",
"url": "https://github.com/voideditor/void/blob/main/extensions/vscode-colorize-tests/test/colorize-fixtures/test.md",
"date": "2024-09-11T02:37:00",
"stars": 10433,
"description": null,
"file_size":... |
*italics*, **bold**, ``literal``.
1. A list
2. With items
- With sub-lists ...
- ... of things.
3. Other things
definition list
A list of terms and their definition
Literal block::
x = 2 + 3
Section separators are all interchangeable.
=====
Title
=====
--------
Subtitle
--------
Section 1
=========... | {
"source": "voideditor/void",
"title": "extensions/vscode-colorize-tests/test/colorize-fixtures/test.rst",
"url": "https://github.com/voideditor/void/blob/main/extensions/vscode-colorize-tests/test/colorize-fixtures/test.rst",
"date": "2024-09-11T02:37:00",
"stars": 10433,
"description": null,
"file_size... |
Tabstops
--
With tabstops you can make the editor cursor move inside a snippet. Use `$1`, `$2` to specify cursor locations. The number is the order in which tabstops will be visited, whereas `$0` denotes the final cursor position. Multiple tabstops are linked and updated in sync.
Placeholders
--
Placeholders are tab... | {
"source": "voideditor/void",
"title": "src/vs/editor/contrib/snippet/browser/snippet.md",
"url": "https://github.com/voideditor/void/blob/main/src/vs/editor/contrib/snippet/browser/snippet.md",
"date": "2024-09-11T02:37:00",
"stars": 10433,
"description": null,
"file_size": 6021
} |
# Diffing Fixture Tests
Every folder in `fixtures` represents a test.
The file that starts with `1.` is diffed against the file that starts with `2.`. Use `tst` instead of `ts` to avoid compiler/linter errors for typescript diff files.
* Missing `*.expected.diff.json` are created automatically (as well as an `*.inval... | {
"source": "voideditor/void",
"title": "src/vs/editor/test/node/diffing/README.md",
"url": "https://github.com/voideditor/void/blob/main/src/vs/editor/test/node/diffing/README.md",
"date": "2024-09-11T02:37:00",
"stars": 10433,
"description": null,
"file_size": 817
} |
Run `node build.js` to compile the React into `out/`.
A couple things to remember:
- Make sure to add .js at the end of any external imports used in here, e.g. ../../../../../my_file.js. If you don't do this, you will get untraceable errors.
- src/ needs to be shallow (1 folder deep) so the detection of externals wo... | {
"source": "voideditor/void",
"title": "src/vs/workbench/contrib/void/browser/react/README.md",
"url": "https://github.com/voideditor/void/blob/main/src/vs/workbench/contrib/void/browser/react/README.md",
"date": "2024-09-11T02:37:00",
"stars": 10433,
"description": null,
"file_size": 354
} |
# F5-TTS: A Fairytaler that Fakes Fluent and Faithful Speech with Flow Matching
[](https://github.com/SWivid/F5-TTS)
[](https://arxiv.org/abs/2410.06885)
[.
2. *LibriSpeech test-clean*: Download from [OpenSLR](http://www.ope... | {
"source": "SWivid/F5-TTS",
"title": "src/f5_tts/eval/README.md",
"url": "https://github.com/SWivid/F5-TTS/blob/main/src/f5_tts/eval/README.md",
"date": "2024-10-08T13:36:55",
"stars": 9872,
"description": "Official code for \"F5-TTS: A Fairytaler that Fakes Fluent and Faithful Speech with Flow Matching\""... |
# Inference
The pretrained model checkpoints can be reached at [🤗 Hugging Face](https://huggingface.co/SWivid/F5-TTS) and [🤖 Model Scope](https://www.modelscope.cn/models/SWivid/F5-TTS_Emilia-ZH-EN), or will be automatically downloaded when running inference scripts.
**More checkpoints with whole community efforts ... | {
"source": "SWivid/F5-TTS",
"title": "src/f5_tts/infer/README.md",
"url": "https://github.com/SWivid/F5-TTS/blob/main/src/f5_tts/infer/README.md",
"date": "2024-10-08T13:36:55",
"stars": 9872,
"description": "Official code for \"F5-TTS: A Fairytaler that Fakes Fluent and Faithful Speech with Flow Matching\... |
<!-- omit in toc -->
# Shared Model Cards
<!-- omit in toc -->
### **Prerequisites of using**
- This document is serving as a quick lookup table for the community training/finetuning result, with various language support.
- The models in this repository are open source and are based on voluntary contributions from con... | {
"source": "SWivid/F5-TTS",
"title": "src/f5_tts/infer/SHARED.md",
"url": "https://github.com/SWivid/F5-TTS/blob/main/src/f5_tts/infer/SHARED.md",
"date": "2024-10-08T13:36:55",
"stars": 9872,
"description": "Official code for \"F5-TTS: A Fairytaler that Fakes Fluent and Faithful Speech with Flow Matching\... |
# Training
## Prepare Dataset
Example data processing scripts, and you may tailor your own one along with a Dataset class in `src/f5_tts/model/dataset.py`.
### 1. Some specific Datasets preparing scripts
Download corresponding dataset first, and fill in the path in scripts.
```bash
# Prepare the Emilia dataset
pyth... | {
"source": "SWivid/F5-TTS",
"title": "src/f5_tts/train/README.md",
"url": "https://github.com/SWivid/F5-TTS/blob/main/src/f5_tts/train/README.md",
"date": "2024-10-08T13:36:55",
"stars": 9872,
"description": "Official code for \"F5-TTS: A Fairytaler that Fakes Fluent and Faithful Speech with Flow Matching\... |
## Backbones quick introduction
### unett.py
- flat unet transformer
- structure same as in e2-tts & voicebox paper except using rotary pos emb
- update: allow possible abs pos emb & convnextv2 blocks for embedded text before concat
### dit.py
- adaln-zero dit
- embedded timestep as condition
- concatted noised_inpu... | {
"source": "SWivid/F5-TTS",
"title": "src/f5_tts/model/backbones/README.md",
"url": "https://github.com/SWivid/F5-TTS/blob/main/src/f5_tts/model/backbones/README.md",
"date": "2024-10-08T13:36:55",
"stars": 9872,
"description": "Official code for \"F5-TTS: A Fairytaler that Fakes Fluent and Faithful Speech... |
# FlashMLA
FlashMLA is an efficient MLA decoding kernel for Hopper GPUs, optimized for variable-length sequences serving.
Currently released:
- BF16, FP16
- Paged kvcache with block size of 64
## Quick start
### Install
```bash
python setup.py install
```
### Benchmark
```bash
python tests/test_flash_mla.py
```
... | {
"source": "deepseek-ai/FlashMLA",
"title": "README.md",
"url": "https://github.com/deepseek-ai/FlashMLA/blob/main/README.md",
"date": "2025-02-21T06:31:27",
"stars": 9433,
"description": null,
"file_size": 1420
} |
# AI Hedge Fund
This is a proof of concept for an AI-powered hedge fund. The goal of this project is to explore the use of AI to make trading decisions. This project is for **educational** purposes only and is not intended for real trading or investment.
This system employs several agents working together:
1. Ben ... | {
"source": "virattt/ai-hedge-fund",
"title": "README.md",
"url": "https://github.com/virattt/ai-hedge-fund/blob/main/README.md",
"date": "2024-11-29T16:30:01",
"stars": 8829,
"description": "An AI Hedge Fund Team",
"file_size": 5993
} |
---
name: Bug report
about: Create a report to help us improve
title: ''
labels: bug
assignees: ''
---
**Describe the bug**
A clear and concise description of what the bug is.
**Screenshot**
Add a screenshot of the bug to help explain your problem.
**Additional context**
Add any other context about the problem here... | {
"source": "virattt/ai-hedge-fund",
"title": ".github/ISSUE_TEMPLATE/bug_report.md",
"url": "https://github.com/virattt/ai-hedge-fund/blob/main/.github/ISSUE_TEMPLATE/bug_report.md",
"date": "2024-11-29T16:30:01",
"stars": 8829,
"description": "An AI Hedge Fund Team",
"file_size": 321
} |
---
name: Feature request
about: Suggest an idea for this project
title: ''
labels: enhancement
assignees: ''
---
**Describe the feature you'd like**
A clear and concise description of what you want to happen. | {
"source": "virattt/ai-hedge-fund",
"title": ".github/ISSUE_TEMPLATE/feature_request.md",
"url": "https://github.com/virattt/ai-hedge-fund/blob/main/.github/ISSUE_TEMPLATE/feature_request.md",
"date": "2024-11-29T16:30:01",
"stars": 8829,
"description": "An AI Hedge Fund Team",
"file_size": 211
} |
## Compilation
Building the project is for users that want to contribute code only.
If you wish to build the emulator yourself, follow these steps:
### Step 1
Install the [.NET 9.0 (or higher) SDK](https://dotnet.microsoft.com/download/dotnet/9.0).
Make sure your SDK version is higher or equal to the required versio... | {
"source": "Ryubing/Ryujinx",
"title": "COMPILING.md",
"url": "https://github.com/Ryubing/Ryujinx/blob/master/COMPILING.md",
"date": "2024-10-06T20:58:11",
"stars": 8135,
"description": "Nintendo Switch emulator written in C#, originally created by gdkchan.",
"file_size": 1073
} |
# Contribution to Ryujinx
You can contribute to Ryujinx with PRs, testing of PRs and issues. Contributing code and other implementations is greatly appreciated alongside simply filing issues for problems you encounter.
Please read the entire document before continuing as it can potentially save everyone involved a sig... | {
"source": "Ryubing/Ryujinx",
"title": "CONTRIBUTING.md",
"url": "https://github.com/Ryubing/Ryujinx/blob/master/CONTRIBUTING.md",
"date": "2024-10-06T20:58:11",
"stars": 8135,
"description": "Nintendo Switch emulator written in C#, originally created by gdkchan.",
"file_size": 9200
} |
<table align="center">
<tr>
<td align="center" width="25%">
<img src="https://raw.githubusercontent.com/GreemDev/ryuassets/refs/heads/main/RyujinxApp_1024.png" alt="Ryujinx" >
</td>
<td align="center" width="75%">
# Ryujinx
[
<details>
<summary>See License</summary>
```
GNU LESSER GENERAL PUBLIC LICENSE
Version 3, 29 June 2007
Copyright (C) 2007 Free Software Foundation, Inc. <http://fsf.org/>
Everyone is permitted to copy and distribute verbatim copies
of this... | {
"source": "Ryubing/Ryujinx",
"title": "distribution/legal/THIRDPARTY.md",
"url": "https://github.com/Ryubing/Ryujinx/blob/master/distribution/legal/THIRDPARTY.md",
"date": "2024-10-06T20:58:11",
"stars": 8135,
"description": "Nintendo Switch emulator written in C#, originally created by gdkchan.",
"file... |
# C# Coding Style
The general rule we follow is "use Visual Studio defaults".
Using an IDE that supports the `.editorconfig` standard will make this much simpler.
1. We use [Allman style](http://en.wikipedia.org/wiki/Indent_style#Allman_style) braces, where each brace begins on a new line. A single line statement bl... | {
"source": "Ryubing/Ryujinx",
"title": "docs/coding-guidelines/coding-style.md",
"url": "https://github.com/Ryubing/Ryujinx/blob/master/docs/coding-guidelines/coding-style.md",
"date": "2024-10-06T20:58:11",
"stars": 8135,
"description": "Nintendo Switch emulator written in C#, originally created by gdkcha... |
# Pull Request Guide
## Contributing Rules
All contributions to GreemDev/Ryujinx repository are made via pull requests (PRs) rather than through direct commits. The pull requests are reviewed and merged by the maintainers after a review and at least two approvals from the core development team.
To merge pull request... | {
"source": "Ryubing/Ryujinx",
"title": "docs/workflow/pr-guide.md",
"url": "https://github.com/Ryubing/Ryujinx/blob/master/docs/workflow/pr-guide.md",
"date": "2024-10-06T20:58:11",
"stars": 8135,
"description": "Nintendo Switch emulator written in C#, originally created by gdkchan.",
"file_size": 4987
} |
<img src="./assets/web-ui.png" alt="Browser Use Web UI" width="full"/>
<br/>
[](https://github.com/browser-use/web-ui/stargazers)
[.
Resources:
- [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/)
- [Microsoft Code of Conduct FAQ](https://opensource.mic... | {
"source": "microsoft/TRELLIS",
"title": "CODE_OF_CONDUCT.md",
"url": "https://github.com/microsoft/TRELLIS/blob/main/CODE_OF_CONDUCT.md",
"date": "2024-12-02T05:44:19",
"stars": 7917,
"description": "Official repo for paper \"Structured 3D Latents for Scalable and Versatile 3D Generation\".",
"file_size... |
# TRELLIS-500K
TRELLIS-500K is a dataset of 500K 3D assets curated from [Objaverse(XL)](https://objaverse.allenai.org/), [ABO](https://amazon-berkeley-objects.s3.amazonaws.com/index.html), [3D-FUTURE](https://tianchi.aliyun.com/specials/promotion/alibaba-3d-future), [HSSD](https://huggingface.co/datasets/hssd/hssd-mod... | {
"source": "microsoft/TRELLIS",
"title": "DATASET.md",
"url": "https://github.com/microsoft/TRELLIS/blob/main/DATASET.md",
"date": "2024-12-02T05:44:19",
"stars": 7917,
"description": "Official repo for paper \"Structured 3D Latents for Scalable and Versatile 3D Generation\".",
"file_size": 8304
} |
<img src="assets/logo.webp" width="100%" align="center">
<h1 align="center">Structured 3D Latents<br>for Scalable and Versatile 3D Generation</h1>
<p align="center"><a href="https://arxiv.org/abs/2412.01506"><img src='https://img.shields.io/badge/arXiv-Paper-red?logo=arxiv&logoColor=white' alt='arXiv'></a>
<a href='htt... | {
"source": "microsoft/TRELLIS",
"title": "README.md",
"url": "https://github.com/microsoft/TRELLIS/blob/main/README.md",
"date": "2024-12-02T05:44:19",
"stars": 7917,
"description": "Official repo for paper \"Structured 3D Latents for Scalable and Versatile 3D Generation\".",
"file_size": 13010
} |
<!-- BEGIN MICROSOFT SECURITY.MD V0.0.9 BLOCK -->
## Security
Microsoft takes the security of our software products and services seriously, which includes all source code repositories managed through our GitHub organizations, which include [Microsoft](https://github.com/Microsoft), [Azure](https://github.com/Azure), ... | {
"source": "microsoft/TRELLIS",
"title": "SECURITY.md",
"url": "https://github.com/microsoft/TRELLIS/blob/main/SECURITY.md",
"date": "2024-12-02T05:44:19",
"stars": 7917,
"description": "Official repo for paper \"Structured 3D Latents for Scalable and Versatile 3D Generation\".",
"file_size": 2655
} |
# TODO: The maintainer of this repo has not yet edited this file
**REPO OWNER**: Do you want Customer Service & Support (CSS) support for this product/project?
- **No CSS support:** Fill out this template with information about how to file issues and get help.
- **Yes CSS support:** Fill out an intake form at [a... | {
"source": "microsoft/TRELLIS",
"title": "SUPPORT.md",
"url": "https://github.com/microsoft/TRELLIS/blob/main/SUPPORT.md",
"date": "2024-12-02T05:44:19",
"stars": 7917,
"description": "Official repo for paper \"Structured 3D Latents for Scalable and Versatile 3D Generation\".",
"file_size": 1242
} |
# Open Source License Attribution
Cosmos uses Open Source components. You can find the details of these open-source projects along with license information below, sorted alphabetically.
We are grateful to the developers for their contributions to open source and acknowledge these below.
## Better-Profanity - [M... | {
"source": "NVIDIA/Cosmos",
"title": "ATTRIBUTIONS.md",
"url": "https://github.com/NVIDIA/Cosmos/blob/main/ATTRIBUTIONS.md",
"date": "2024-12-30T17:21:14",
"stars": 7578,
"description": "Cosmos is a world model development platform that consists of world foundation models, tokenizers and video processing p... |
# How to Contribute
We'd love to receive your patches and contributions. Please keep your PRs as draft until such time that you would like us to review them.
## Code Reviews
All submissions, including submissions by project members, require review. We use GitHub pull requests for this purpose. Consult
[GitHub Help](... | {
"source": "NVIDIA/Cosmos",
"title": "CONTRIBUTING.md",
"url": "https://github.com/NVIDIA/Cosmos/blob/main/CONTRIBUTING.md",
"date": "2024-12-30T17:21:14",
"stars": 7578,
"description": "Cosmos is a world model development platform that consists of world foundation models, tokenizers and video processing p... |
# Cosmos Installation
We have only tested the installation with Ubuntu 24.04, 22.04, and 20.04.
1. Install the [NVIDIA Container Toolkit](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/install-guide.html).
2. Clone the repository.
```bash
git clone git@github.com:NVIDIA/Cosmos.git
cd Cosmo... | {
"source": "NVIDIA/Cosmos",
"title": "INSTALL.md",
"url": "https://github.com/NVIDIA/Cosmos/blob/main/INSTALL.md",
"date": "2024-12-30T17:21:14",
"stars": 7578,
"description": "Cosmos is a world model development platform that consists of world foundation models, tokenizers and video processing pipeline to... |

--------------------------------------------------------------------------------
### [Website](https://www.nvidia.com/en-us/ai/cosmos/) | [HuggingFace](https://huggingface.co/collections/nvidia/cosmos-6751e884dc10e013a0a0d8e6) | [GPU-free Preview](https://build.nvidia.com/explore... | {
"source": "NVIDIA/Cosmos",
"title": "README.md",
"url": "https://github.com/NVIDIA/Cosmos/blob/main/README.md",
"date": "2024-12-30T17:21:14",
"stars": 7578,
"description": "Cosmos is a world model development platform that consists of world foundation models, tokenizers and video processing pipeline to a... |
# Release Cadence
| Version | Description | Date |
|------------|----------|----------|
| [v1.0](release_notes/v0p1.md) | Initial diffusion and autoregressive WFMs release | 2025-01-06 |
| [v0.1](release_notes/v0p1.md) | Initial tokenizer release | 2024-11-06 | | {
"source": "NVIDIA/Cosmos",
"title": "RELEASE.md",
"url": "https://github.com/NVIDIA/Cosmos/blob/main/RELEASE.md",
"date": "2024-12-30T17:21:14",
"stars": 7578,
"description": "Cosmos is a world model development platform that consists of world foundation models, tokenizers and video processing pipeline to... |
# Checkpoint directory
Follow our instructions for downloading checkpoints in [Cosmos Diffusion Inference](../cosmos1/models/diffusion/README.md#download-checkpoints) and [Cosmos Autoregressive Inference](../cosmos1/models/autoregressive/README.md). Cosmos checkpoints will be downloaded to this directory. | {
"source": "NVIDIA/Cosmos",
"title": "checkpoints/README.md",
"url": "https://github.com/NVIDIA/Cosmos/blob/main/checkpoints/README.md",
"date": "2024-12-30T17:21:14",
"stars": 7578,
"description": "Cosmos is a world model development platform that consists of world foundation models, tokenizers and video ... |
# Release note
- Cosmos 0.1 was released with the [Cosmos Tokenizer Webage](https://research.nvidia.com/labs/dir/cosmos-tokenizer/).
- 10 tokenizers were released in the [Hugging Face](https://huggingface.co/collections/nvidia/cosmos-6751e884dc10e013a0a0d8e6) as shown in the table below.
- Inference scripts for the mo... | {
"source": "NVIDIA/Cosmos",
"title": "release_notes/v0p1.md",
"url": "https://github.com/NVIDIA/Cosmos/blob/main/release_notes/v0p1.md",
"date": "2024-12-30T17:21:14",
"stars": 7578,
"description": "Cosmos is a world model development platform that consists of world foundation models, tokenizers and video ... |
# Release Notes
# [02/10/2025](https://github.com/NVIDIA/Cosmos/commit/868ff171b9d676c53e094c4324a45a5f06d749e2)
- Cosmos Tokenizer inference and post-training support
- Cosmos Video2World post-training support
# [01/27/2025](https://github.com/NVIDIA/Cosmos/commit/c82c9dc6f9a2f046033d0a26ec525bc389b641ef)
- Stabilit... | {
"source": "NVIDIA/Cosmos",
"title": "release_notes/v1p0.md",
"url": "https://github.com/NVIDIA/Cosmos/blob/main/release_notes/v1p0.md",
"date": "2024-12-30T17:21:14",
"stars": 7578,
"description": "Cosmos is a world model development platform that consists of world foundation models, tokenizers and video ... |
# Cosmos Post-training
In the [Cosmos paper](https://research.nvidia.com/publication/2025-01_cosmos-world-foundation-model-platform-physical-ai), we discuss several post-training examples of Cosmos pre-trained World Foundation Models (WFMs) for various Physical AI tasks, including
- General Post-Training: Fine-tune t... | {
"source": "NVIDIA/Cosmos",
"title": "cosmos1/models/POST_TRAINING.md",
"url": "https://github.com/NVIDIA/Cosmos/blob/main/cosmos1/models/POST_TRAINING.md",
"date": "2024-12-30T17:21:14",
"stars": 7578,
"description": "Cosmos is a world model development platform that consists of world foundation models, t... |
# Cosmos Autoregressive-based World Foundation Models
## Table of Contents
- [Getting Started](#getting-started)
- [Set Up Docker Environment](#set-up-docker-environment)
- [Download Checkpoints](#download-checkpoints)
- [Usage](#usage)
- [Model Types](#model-types)
- [Single and Batch Generation](#single-and-... | {
"source": "NVIDIA/Cosmos",
"title": "cosmos1/models/autoregressive/README.md",
"url": "https://github.com/NVIDIA/Cosmos/blob/main/cosmos1/models/autoregressive/README.md",
"date": "2024-12-30T17:21:14",
"stars": 7578,
"description": "Cosmos is a world model development platform that consists of world foun... |
# Cosmos Diffusion-based World Foundation Models
## Table of Contents
- [Getting Started](#getting-started)
- [Set Up Docker Environment](#set-up-docker-environment)
- [Download Checkpoints](#download-checkpoints)
- [Usage](#usage)
- [Model Types](#model-types)
- [Single and Batch Generation](#single-and-batch... | {
"source": "NVIDIA/Cosmos",
"title": "cosmos1/models/diffusion/README.md",
"url": "https://github.com/NVIDIA/Cosmos/blob/main/cosmos1/models/diffusion/README.md",
"date": "2024-12-30T17:21:14",
"stars": 7578,
"description": "Cosmos is a world model development platform that consists of world foundation mod... |
# Cosmos Guardrail
This page outlines a set of tools to ensure content safety in Cosmos. For implementation details, please consult the [Cosmos paper](https://research.nvidia.com/publication/2025-01_cosmos-world-foundation-model-platform-physical-ai).
## Overview
Our guardrail system consists of two stages: pre-Guar... | {
"source": "NVIDIA/Cosmos",
"title": "cosmos1/models/guardrail/README.md",
"url": "https://github.com/NVIDIA/Cosmos/blob/main/cosmos1/models/guardrail/README.md",
"date": "2024-12-30T17:21:14",
"stars": 7578,
"description": "Cosmos is a world model development platform that consists of world foundation mod... |
<!-- # SPDX-FileCopyrightText: Copyright (c) 2025 NVIDIA CORPORATION & AFFILIATES. All rights reserved.
# SPDX-License-Identifier: Apache-2.0
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#... | {
"source": "NVIDIA/Cosmos",
"title": "cosmos1/models/tokenizer/README.md",
"url": "https://github.com/NVIDIA/Cosmos/blob/main/cosmos1/models/tokenizer/README.md",
"date": "2024-12-30T17:21:14",
"stars": 7578,
"description": "Cosmos is a world model development platform that consists of world foundation mod... |
# Cosmos Tokenizer: NeMo Framework Finetuning User Guide
Post-train the Cosmos Tokenizer using the [NVIDIA NeMo Framework](https://docs.nvidia.com/nemo-framework/user-guide/latest/overview.html) to more accurately model previously unseen scenarios in your customer data, particularly for self-driving applications. By a... | {
"source": "NVIDIA/Cosmos",
"title": "cosmos1/models/tokenizer/nemo/README.md",
"url": "https://github.com/NVIDIA/Cosmos/blob/main/cosmos1/models/tokenizer/nemo/README.md",
"date": "2024-12-30T17:21:14",
"stars": 7578,
"description": "Cosmos is a world model development platform that consists of world foun... |
# Cosmos Autoregressive-based World Foundation Models: NeMo Framework User Guide
Learn how to [run inference](#run-inference) with Cosmos Autoregressive-based World Foundation Models (WFMs) using the [NVIDIA NeMo Framework](https://docs.nvidia.com/nemo-framework/user-guide/latest/overview.html) for your custom Physica... | {
"source": "NVIDIA/Cosmos",
"title": "cosmos1/models/autoregressive/nemo/inference/README.md",
"url": "https://github.com/NVIDIA/Cosmos/blob/main/cosmos1/models/autoregressive/nemo/inference/README.md",
"date": "2024-12-30T17:21:14",
"stars": 7578,
"description": "Cosmos is a world model development platfo... |
# Cosmos Autoregressive-based World Foundation Models: NeMo Framework User Guide
Learn how to [post-train](#post-train) Cosmos Autoregressive-based World Foundation Models (WFMs) using the [NVIDIA NeMo Framework](https://docs.nvidia.com/nemo-framework/user-guide/latest/overview.html) for your custom Physical AI tasks ... | {
"source": "NVIDIA/Cosmos",
"title": "cosmos1/models/autoregressive/nemo/post_training/README.md",
"url": "https://github.com/NVIDIA/Cosmos/blob/main/cosmos1/models/autoregressive/nemo/post_training/README.md",
"date": "2024-12-30T17:21:14",
"stars": 7578,
"description": "Cosmos is a world model developmen... |
# Cosmos Diffusion-based World Foundation Models: NeMo Framework User Guide
Learn how to [run inference](#inference) with Cosmos Diffusion-based World Foundation Models (WFMs) using the [NVIDIA NeMo Framework](https://docs.nvidia.com/nemo-framework/user-guide/latest/overview.html) for your custom Physical AI tasks by ... | {
"source": "NVIDIA/Cosmos",
"title": "cosmos1/models/diffusion/nemo/inference/README.md",
"url": "https://github.com/NVIDIA/Cosmos/blob/main/cosmos1/models/diffusion/nemo/inference/README.md",
"date": "2024-12-30T17:21:14",
"stars": 7578,
"description": "Cosmos is a world model development platform that co... |
# Cosmos Diffusion-based World Foundation Models: NeMo Framework User Guide
Learn how to [post-train](#post-train) Cosmos Diffusion-based World Foundation Models (WFMs) using the [NVIDIA NeMo Framework](https://docs.nvidia.com/nemo-framework/user-guide/latest/overview.html) for your custom Physical AI tasks by followi... | {
"source": "NVIDIA/Cosmos",
"title": "cosmos1/models/diffusion/nemo/post_training/README.md",
"url": "https://github.com/NVIDIA/Cosmos/blob/main/cosmos1/models/diffusion/nemo/post_training/README.md",
"date": "2024-12-30T17:21:14",
"stars": 7578,
"description": "Cosmos is a world model development platform... |
# Contributor Covenant Code of Conduct
## Our Pledge
We as members, contributors, and leaders pledge to make participation in our
community a harassment-free experience for everyone, regardless of age, body
size, visible or invisible disability, ethnicity, sex characteristics, gender
identity and expression, level of... | {
"source": "cyclotruc/gitingest",
"title": "CODE_OF_CONDUCT.md",
"url": "https://github.com/cyclotruc/gitingest/blob/main/CODE_OF_CONDUCT.md",
"date": "2024-11-29T08:27:18",
"stars": 6697,
"description": "Replace 'hub' with 'ingest' in any github url to get a prompt-friendly extract of a codebase ",
"fil... |
# Contributing to Gitingest
Thanks for your interest in contributing to Gitingest! 🚀 Gitingest aims to be friendly for first time contributors, with a simple Python and HTML codebase. We would love your help to make it even better. If you need any help while working with the code, please reach out to us on [Discord](... | {
"source": "cyclotruc/gitingest",
"title": "CONTRIBUTING.md",
"url": "https://github.com/cyclotruc/gitingest/blob/main/CONTRIBUTING.md",
"date": "2024-11-29T08:27:18",
"stars": 6697,
"description": "Replace 'hub' with 'ingest' in any github url to get a prompt-friendly extract of a codebase ",
"file_size... |
# Gitingest
[](https://gitingest.com)
[](https://github.com/cyclotruc/gitingest/blob/main/LICENSE)
[](https://badge.fury.io/py/gitingest)
[, [Hsiang-Wei H... | {
"source": "yangchris11/samurai",
"title": "README.md",
"url": "https://github.com/yangchris11/samurai/blob/master/README.md",
"date": "2024-11-06T22:46:05",
"stars": 6541,
"description": "Official repository of \"SAMURAI: Adapting Segment Anything Model for Zero-Shot Visual Tracking with Motion-Aware Memo... |
# Code of Conduct
## Our Pledge
In the interest of fostering an open and welcoming environment, we as
contributors and maintainers pledge to make participation in our project and
our community a harassment-free experience for everyone, regardless of age, body
size, disability, ethnicity, sex characteristics, gender i... | {
"source": "yangchris11/samurai",
"title": "sam2/CODE_OF_CONDUCT.md",
"url": "https://github.com/yangchris11/samurai/blob/master/sam2/CODE_OF_CONDUCT.md",
"date": "2024-11-06T22:46:05",
"stars": 6541,
"description": "Official repository of \"SAMURAI: Adapting Segment Anything Model for Zero-Shot Visual Tra... |
# Contributing to segment-anything
We want to make contributing to this project as easy and transparent as
possible.
## Pull Requests
We actively welcome your pull requests.
1. Fork the repo and create your branch from `main`.
2. If you've added code that should be tested, add tests.
3. If you've changed APIs, update... | {
"source": "yangchris11/samurai",
"title": "sam2/CONTRIBUTING.md",
"url": "https://github.com/yangchris11/samurai/blob/master/sam2/CONTRIBUTING.md",
"date": "2024-11-06T22:46:05",
"stars": 6541,
"description": "Official repository of \"SAMURAI: Adapting Segment Anything Model for Zero-Shot Visual Tracking ... |
## Installation
### Requirements
- Linux with Python ≥ 3.10, PyTorch ≥ 2.3.1 and [torchvision](https://github.com/pytorch/vision/) that matches the PyTorch installation. Install them together at https://pytorch.org to ensure this.
* Note older versions of Python or PyTorch may also work. However, the versions above... | {
"source": "yangchris11/samurai",
"title": "sam2/INSTALL.md",
"url": "https://github.com/yangchris11/samurai/blob/master/sam2/INSTALL.md",
"date": "2024-11-06T22:46:05",
"stars": 6541,
"description": "Official repository of \"SAMURAI: Adapting Segment Anything Model for Zero-Shot Visual Tracking with Motio... |
# SAM 2: Segment Anything in Images and Videos
**[AI at Meta, FAIR](https://ai.meta.com/research/)**
[Nikhila Ravi](https://nikhilaravi.com/), [Valentin Gabeur](https://gabeur.github.io/), [Yuan-Ting Hu](https://scholar.google.com/citations?user=E8DVVYQAAAAJ&hl=en), [Ronghang Hu](https://ronghanghu.com/), [Chaitanya ... | {
"source": "yangchris11/samurai",
"title": "sam2/README.md",
"url": "https://github.com/yangchris11/samurai/blob/master/sam2/README.md",
"date": "2024-11-06T22:46:05",
"stars": 6541,
"description": "Official repository of \"SAMURAI: Adapting Segment Anything Model for Zero-Shot Visual Tracking with Motion-... |
# SAM 2 Demo
Welcome to the SAM 2 Demo! This project consists of a frontend built with React TypeScript and Vite and a backend service using Python Flask and Strawberry GraphQL. Both components can be run in Docker containers or locally on MPS (Metal Performance Shaders) or CPU. However, running the backend service on... | {
"source": "yangchris11/samurai",
"title": "sam2/demo/README.md",
"url": "https://github.com/yangchris11/samurai/blob/master/sam2/demo/README.md",
"date": "2024-11-06T22:46:05",
"stars": 6541,
"description": "Official repository of \"SAMURAI: Adapting Segment Anything Model for Zero-Shot Visual Tracking wi... |
# Segment Anything Video (SA-V) Dataset
## Overview
[Segment Anything Video (SA-V)](https://ai.meta.com/datasets/segment-anything-video/), consists of 51K diverse videos and 643K high-quality spatio-temporal segmentation masks (i.e., masklets). The dataset is released under the CC by 4.0 license. Browse the dataset [... | {
"source": "yangchris11/samurai",
"title": "sam2/sav_dataset/README.md",
"url": "https://github.com/yangchris11/samurai/blob/master/sam2/sav_dataset/README.md",
"date": "2024-11-06T22:46:05",
"stars": 6541,
"description": "Official repository of \"SAMURAI: Adapting Segment Anything Model for Zero-Shot Visu... |
## SAM 2 toolkits
This directory provides toolkits for additional SAM 2 use cases.
### Semi-supervised VOS inference
The `vos_inference.py` script can be used to generate predictions for semi-supervised video object segmentation (VOS) evaluation on datasets such as [DAVIS](https://davischallenge.org/index.html), [MO... | {
"source": "yangchris11/samurai",
"title": "sam2/tools/README.md",
"url": "https://github.com/yangchris11/samurai/blob/master/sam2/tools/README.md",
"date": "2024-11-06T22:46:05",
"stars": 6541,
"description": "Official repository of \"SAMURAI: Adapting Segment Anything Model for Zero-Shot Visual Tracking ... |
# Training Code for SAM 2
This folder contains the training code for SAM 2, a foundation model for promptable visual segmentation in images and videos.
The code allows users to train and fine-tune SAM 2 on their own datasets (image, video, or both).
## Structure
The training code is organized into the following sub... | {
"source": "yangchris11/samurai",
"title": "sam2/training/README.md",
"url": "https://github.com/yangchris11/samurai/blob/master/sam2/training/README.md",
"date": "2024-11-06T22:46:05",
"stars": 6541,
"description": "Official repository of \"SAMURAI: Adapting Segment Anything Model for Zero-Shot Visual Tra... |
# README
## Description for different text files
GOT10K
- got10k_train_full_split.txt: the complete GOT-10K training set. (9335 videos)
- got10k_train_split.txt: part of videos from the GOT-10K training set
- got10k_val_split.txt: another part of videos from the GOT-10K training set
- got10k_vot_exclude.txt: 1k videos... | {
"source": "yangchris11/samurai",
"title": "lib/train/data_specs/README.md",
"url": "https://github.com/yangchris11/samurai/blob/master/lib/train/data_specs/README.md",
"date": "2024-11-06T22:46:05",
"stars": 6541,
"description": "Official repository of \"SAMURAI: Adapting Segment Anything Model for Zero-S... |
# Awesome LLM Strawberry (OpenAI o1)
[](https://github.com/sindresorhus/awesome)  
(中文 | [English](README_en.md) | [日本語](README_ja.md))
这是虾哥的第一个硬件作品。
👉 [ESP32+SenseVoice+Qwen72B打造你的AI聊天伴侣!【bilibili】](https://www.bilibili.com/video/BV11msTenEH3/)
👉 [给小智装上 DeepSeek 的聪明大脑【bilibili】](https://www.bilibili.com/video/BV1GQP6eNEFG/)
👉 [手工打造你的 AI 女友,新手入门教程【bilibili】]... | {
"source": "78/xiaozhi-esp32",
"title": "README.md",
"url": "https://github.com/78/xiaozhi-esp32/blob/main/README.md",
"date": "2024-08-31T10:08:16",
"stars": 6444,
"description": "Build your own AI friend",
"file_size": 5587
} |
# XiaoZhi AI Chatbot
([中文](README.md) | English | [日本語](README_ja.md))
This is Terrence's first hardware project.
👉 [Build your AI chat companion with ESP32+SenseVoice+Qwen72B!【bilibili】](https://www.bilibili.com/video/BV11msTenEH3/)
👉 [Equipping XiaoZhi with DeepSeek's smart brain【bilibili】](https://www.bilibili... | {
"source": "78/xiaozhi-esp32",
"title": "README_en.md",
"url": "https://github.com/78/xiaozhi-esp32/blob/main/README_en.md",
"date": "2024-08-31T10:08:16",
"stars": 6444,
"description": "Build your own AI friend",
"file_size": 6983
} |
# シャオジー AI チャットボット
([中文](README.md) | [English](README_en.md) | 日本語)
これは シャーガー(Terrence)の最初のハードウェア作品です。
👉 [ESP32+SenseVoice+Qwen72Bで AI チャット仲間を作ろう!【bilibili】](https://www.bilibili.com/video/BV11msTenEH3/)
👉 [シャオジーに DeepSeek のスマートな頭脳を搭載【bilibili】](https://www.bilibili.com/video/BV1GQP6eNEFG/)
👉 [自分だけの AI パートナーを作... | {
"source": "78/xiaozhi-esp32",
"title": "README_ja.md",
"url": "https://github.com/78/xiaozhi-esp32/blob/main/README_ja.md",
"date": "2024-08-31T10:08:16",
"stars": 6444,
"description": "Build your own AI friend",
"file_size": 5480
} |
以下是一份基于代码实现整理的 WebSocket 通信协议文档,概述客户端(设备)与服务器之间如何通过 WebSocket 进行交互。该文档仅基于所提供的代码推断,实际部署时可能需要结合服务器端实现进行进一步确认或补充。
---
## 1. 总体流程概览
1. **设备端初始化**
- 设备上电、初始化 `Application`:
- 初始化音频编解码器、显示屏、LED 等
- 连接网络
- 创建并初始化实现 `Protocol` 接口的 WebSocket 协议实例(`WebsocketProtocol`)
- 进入主循环等待事件(音频输入、音频输出、调度任务等... | {
"source": "78/xiaozhi-esp32",
"title": "docs/websocket.md",
"url": "https://github.com/78/xiaozhi-esp32/blob/main/docs/websocket.md",
"date": "2024-08-31T10:08:16",
"stars": 6444,
"description": "Build your own AI friend",
"file_size": 7593
} |
# 编译配置命令
**配置编译目标为 ESP32:**
```bash
idf.py set-target esp32
```
**打开 menuconfig:**
```bash
idf.py menuconfig
```
**选择板子:**
```
Xiaozhi Assistant -> Board Type -> AtomMatrix + Echo Base
```
**修改 flash 大小:**
```
Serial flasher config -> Flash size -> 4 MB
```
**修改分区表:**
```
Partition Table -> Custom partition C... | {
"source": "78/xiaozhi-esp32",
"title": "main/boards/atommatrix-echo-base/README.md",
"url": "https://github.com/78/xiaozhi-esp32/blob/main/main/boards/atommatrix-echo-base/README.md",
"date": "2024-08-31T10:08:16",
"stars": 6444,
"description": "Build your own AI friend",
"file_size": 387
} |
# 编译配置命令
**配置编译目标为 ESP32S3:**
```bash
idf.py set-target esp32s3
```
**打开 menuconfig:**
```bash
idf.py menuconfig
```
**选择板子:**
```
Xiaozhi Assistant -> Board Type -> AtomS3 + Echo Base
```
**关闭语音唤醒:**
```
Xiaozhi Assistant -> [ ] 启用语音唤醒与音频处理 -> Unselect
```
**修改 flash 大小:**
```
Serial flasher config -> Flash ... | {
"source": "78/xiaozhi-esp32",
"title": "main/boards/atoms3-echo-base/README.md",
"url": "https://github.com/78/xiaozhi-esp32/blob/main/main/boards/atoms3-echo-base/README.md",
"date": "2024-08-31T10:08:16",
"stars": 6444,
"description": "Build your own AI friend",
"file_size": 573
} |
# 编译配置命令
**配置编译目标为 ESP32S3:**
```bash
idf.py set-target esp32s3
```
**打开 menuconfig:**
```bash
idf.py menuconfig
```
**选择板子:**
```
Xiaozhi Assistant -> Board Type -> AtomS3R + Echo Base
```
**修改 flash 大小:**
```
Serial flasher config -> Flash size -> 8 MB
```
**修改分区表:**
```
Partiti... | {
"source": "78/xiaozhi-esp32",
"title": "main/boards/atoms3r-echo-base/README.md",
"url": "https://github.com/78/xiaozhi-esp32/blob/main/main/boards/atoms3r-echo-base/README.md",
"date": "2024-08-31T10:08:16",
"stars": 6444,
"description": "Build your own AI friend",
"file_size": 544
} |
# 编译配置命令
**配置编译目标为 ESP32:**
```bash
idf.py set-target esp32
```
**打开 menuconfig:**
```bash
idf.py menuconfig
```
**选择板子:**
```
Xiaozhi Assistant -> Board Type -> 面包板 ESP32 DevKit
```
**修改 flash 大小:**
```
Serial flasher config -> Flash size -> 4 MB
```
**修改分区表:**
```
Partition Table -> Custom partition CSV fil... | {
"source": "78/xiaozhi-esp32",
"title": "main/boards/bread-compact-esp32/README.md",
"url": "https://github.com/78/xiaozhi-esp32/blob/main/main/boards/bread-compact-esp32/README.md",
"date": "2024-08-31T10:08:16",
"stars": 6444,
"description": "Build your own AI friend",
"file_size": 381
} |
## 立创·实战派ESP32-C3开发板
1、开发板资料:https://wiki.lckfb.com/zh-hans/szpi-esp32c3
2、该开发板 flash 大小为 8MB,编译时注意选择合适的分区表:
```
Partition Table --->
Partition Table (Custom partition table CSV) --->
(partitions_8M.csv) Custom partition CSV file
``` | {
"source": "78/xiaozhi-esp32",
"title": "main/boards/lichuang-c3-dev/README.md",
"url": "https://github.com/78/xiaozhi-esp32/blob/main/main/boards/lichuang-c3-dev/README.md",
"date": "2024-08-31T10:08:16",
"stars": 6444,
"description": "Build your own AI friend",
"file_size": 242
} |
# 编译配置命令
**配置编译目标为 ESP32S3:**
```bash
idf.py set-target esp32s3
```
**打开 menuconfig:**
```bash
idf.py menuconfig
```
**选择板子:**
```
Xiaozhi Assistant -> Board Type -> M5Stack CoreS3
```
**修改 psram 配置:**
```
Component config -> ESP PSRAM -> SPI RAM config -> Mode (QUAD/OCT) -> Quad Mode PSR... | {
"source": "78/xiaozhi-esp32",
"title": "main/boards/m5stack-core-s3/README.md",
"url": "https://github.com/78/xiaozhi-esp32/blob/main/main/boards/m5stack-core-s3/README.md",
"date": "2024-08-31T10:08:16",
"stars": 6444,
"description": "Build your own AI friend",
"file_size": 368
} |
# 编译配置命令
**配置编译目标为 ESP32S3:**
```bash
idf.py set-target esp32s3
```
**打开 menuconfig:**
```bash
idf.py menuconfig
```
**选择板子:**
```
Xiaozhi Assistant -> Board Type -> Movecall Moji 小智AI衍生版
```
**编译:**
```bash
idf.py build
``` | {
"source": "78/xiaozhi-esp32",
"title": "main/boards/movecall-moji-esp32s3/README.md",
"url": "https://github.com/78/xiaozhi-esp32/blob/main/main/boards/movecall-moji-esp32s3/README.md",
"date": "2024-08-31T10:08:16",
"stars": 6444,
"description": "Build your own AI friend",
"file_size": 258
} |
# 编译配置命令
**配置编译目标为 ESP32S3:**
```bash
idf.py set-target esp32s3
```
**打开 menuconfig:**
```bash
idf.py menuconfig
```
**选择板子:**
```
Xiaozhi Assistant -> Board Type -> 太极小派esp32s3
```
**编译:**
```bash
idf.py build
``` | {
"source": "78/xiaozhi-esp32",
"title": "main/boards/taiji-pi-s3/README.md",
"url": "https://github.com/78/xiaozhi-esp32/blob/main/main/boards/taiji-pi-s3/README.md",
"date": "2024-08-31T10:08:16",
"stars": 6444,
"description": "Build your own AI friend",
"file_size": 222
} |

# Cofounder | Early alpha release
* project - [cofounder.openinterface.ai](https://cofounder.openinterface.ai)
* 👋 [@n_raidenai](https://x.com/n_raidenai)
**cofounder**
- full stack generative web apps ; backend +... | {
"source": "raidendotai/cofounder",
"title": "README.md",
"url": "https://github.com/raidendotai/cofounder/blob/main/README.md",
"date": "2024-09-19T00:27:55",
"stars": 6271,
"description": "ai-generated apps , full stack + generative UI",
"file_size": 7284
} |
A non-ordered roadmap & todo dump
will update with proper map later, ignore for now
---
## nearest
merge with browser-based local dev env using webcontainers ; console.cofounder.openinterface.ai
## validation, errorfix
post-generation validation swarm modules
swarm autofix modules, merge
babel parse
## build, deplo... | {
"source": "raidendotai/cofounder",
"title": "TODO.md",
"url": "https://github.com/raidendotai/cofounder/blob/main/TODO.md",
"date": "2024-09-19T00:27:55",
"stars": 6271,
"description": "ai-generated apps , full stack + generative UI",
"file_size": 1271
} |
## How to start apps
Your backend & vite+react web app will incrementally generate inside `./apps/{YourApp}`
Open your terminal in `./apps/{YourApp}` and run
```sh
npm i && npm run dev
``` | {
"source": "raidendotai/cofounder",
"title": "apps/README.md",
"url": "https://github.com/raidendotai/cofounder/blob/main/apps/README.md",
"date": "2024-09-19T00:27:55",
"stars": 6271,
"description": "ai-generated apps , full stack + generative UI",
"file_size": 190
} |
app generated from cofounder/boilerplate
instructions here on how to start api and frontend , whether in parallel or separately | {
"source": "raidendotai/cofounder",
"title": "cofounder/boilerplate/README.md",
"url": "https://github.com/raidendotai/cofounder/blob/main/cofounder/boilerplate/README.md",
"date": "2024-09-19T00:27:55",
"stars": 6271,
"description": "ai-generated apps , full stack + generative UI",
"file_size": 128
} |
# React + TypeScript + Vite
This template provides a minimal setup to get React working in Vite with HMR and some ESLint rules.
Currently, two official plugins are available:
- [@vitejs/plugin-react](https://github.com/vitejs/vite-plugin-react/blob/main/packages/plugin-react/README.md) uses [Babel](https://babeljs.i... | {
"source": "raidendotai/cofounder",
"title": "cofounder/boilerplate/vitereact-boilerplate/README.md",
"url": "https://github.com/raidendotai/cofounder/blob/main/cofounder/boilerplate/vitereact-boilerplate/README.md",
"date": "2024-09-19T00:27:55",
"stars": 6271,
"description": "ai-generated apps , full sta... |
## Notes
This is a demo design system and will be replaced on official post-alpha release
## Credits
Renders dumped from Figma Presets:
- Blocks.pm by Hexa Plugin
- Google Material & Figma Core Design systems | {
"source": "raidendotai/cofounder",
"title": "cofounder/api/system/presets/ui/design/systems/protoboy-v1/README.md",
"url": "https://github.com/raidendotai/cofounder/blob/main/cofounder/api/system/presets/ui/design/systems/protoboy-v1/README.md",
"date": "2024-09-19T00:27:55",
"stars": 6271,
"description":... |
<div align="center">
<img src="app-icon.png" alt="NeoHtop Logo" width="120" />
<h1>NeoHtop</h1>
<p>A modern, cross-platform system monitor built on top of Svelte, Rust, and Tauri.</p>
[](https://github.com/Abdenasser/neohtop/blob/main/LICENSE)... | {
"source": "Abdenasser/neohtop",
"title": "README.md",
"url": "https://github.com/Abdenasser/neohtop/blob/main/README.md",
"date": "2024-10-30T20:44:11",
"stars": 5821,
"description": "💪🏻 Blazing-fast system monitoring for your desktop (built with Rust, Tauri & Svelte)",
"file_size": 4566
} |
# Contributing to NeoHtop
Thank you for considering contributing to NeoHtop! We welcome contributions from the community.
## How to Contribute
1. Fork the repository.
2. Create a new branch (`git checkout -b feature/YourFeature`).
3. Make your changes.
4. Commit your changes (`git commit -m 'Add some feature'`).
5. ... | {
"source": "Abdenasser/neohtop",
"title": ".github/CONTRIBUTING.md",
"url": "https://github.com/Abdenasser/neohtop/blob/main/.github/CONTRIBUTING.md",
"date": "2024-10-30T20:44:11",
"stars": 5821,
"description": "💪🏻 Blazing-fast system monitoring for your desktop (built with Rust, Tauri & Svelte)",
"fi... |
## Description
Please include a summary of the changes and the related issue. Please also include relevant motivation and context.
Fixes # (issue)
## Type of change
Please delete options that are not relevant.
- [ ] Bug fix (non-breaking change which fixes an issue)
- [ ] New feature (non-breaking change which add... | {
"source": "Abdenasser/neohtop",
"title": ".github/pull_request_template.md",
"url": "https://github.com/Abdenasser/neohtop/blob/main/.github/pull_request_template.md",
"date": "2024-10-30T20:44:11",
"stars": 5821,
"description": "💪🏻 Blazing-fast system monitoring for your desktop (built with Rust, Tauri... |
---
name: Bug report
about: Create a report to help us improve
title: ''
labels: bug
assignees: ''
---
**Describe the bug**
A clear and concise description of what the bug is.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
**Expected behavio... | {
"source": "Abdenasser/neohtop",
"title": ".github/ISSUE_TEMPLATE/bug_report.md",
"url": "https://github.com/Abdenasser/neohtop/blob/main/.github/ISSUE_TEMPLATE/bug_report.md",
"date": "2024-10-30T20:44:11",
"stars": 5821,
"description": "💪🏻 Blazing-fast system monitoring for your desktop (built with Rus... |
---
name: Feature request
about: Suggest an idea for this project
title: ''
labels: enhancement
assignees: ''
---
**Is your feature request related to a problem? Please describe.**
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
**Describe the solution you'd like**
A clea... | {
"source": "Abdenasser/neohtop",
"title": ".github/ISSUE_TEMPLATE/feature_request.md",
"url": "https://github.com/Abdenasser/neohtop/blob/main/.github/ISSUE_TEMPLATE/feature_request.md",
"date": "2024-10-30T20:44:11",
"stars": 5821,
"description": "💪🏻 Blazing-fast system monitoring for your desktop (buil... |
<div align="center">
<h1>s1: Simple test-time scaling</h1>
<p>Minimal recipe for test-time scaling and strong reasoning performance matching o1-preview with just 1,000 examples & budget forcing
</p>
</div>
<br>

****************************************************************
**Updates:*... | {
"source": "simplescaling/s1",
"title": "README.md",
"url": "https://github.com/simplescaling/s1/blob/main/README.md",
"date": "2025-02-01T02:38:16",
"stars": 5696,
"description": "s1: Simple test-time scaling",
"file_size": 11028
} |
MIT License
Copyright (c) 2020 EleutherAI
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distri... | {
"source": "simplescaling/s1",
"title": "eval/lm-evaluation-harness/LICENSE.md",
"url": "https://github.com/simplescaling/s1/blob/main/eval/lm-evaluation-harness/LICENSE.md",
"date": "2025-02-01T02:38:16",
"stars": 5696,
"description": "s1: Simple test-time scaling",
"file_size": 1066
} |
# Language Model Evaluation Harness
[](https://doi.org/10.5281/zenodo.10256836)
---
*Latest News 📣*
- [2024/09] We are prototyping allowing users of LM Evaluation Harness to create and evaluate on text+image multimodal input, text output tasks, and ha... | {
"source": "simplescaling/s1",
"title": "eval/lm-evaluation-harness/README.md",
"url": "https://github.com/simplescaling/s1/blob/main/eval/lm-evaluation-harness/README.md",
"date": "2025-02-01T02:38:16",
"stars": 5696,
"description": "s1: Simple test-time scaling",
"file_size": 39773
} |
# TemplateAPI Usage Guide
The `TemplateAPI` class is a versatile superclass designed to facilitate the integration of various API-based language models into the lm-evaluation-harness framework. This guide will explain how to use and extend the `TemplateAPI` class to implement your own API models. If your API implement... | {
"source": "simplescaling/s1",
"title": "eval/lm-evaluation-harness/docs/API_guide.md",
"url": "https://github.com/simplescaling/s1/blob/main/eval/lm-evaluation-harness/docs/API_guide.md",
"date": "2025-02-01T02:38:16",
"stars": 5696,
"description": "s1: Simple test-time scaling",
"file_size": 7742
} |
# Contributing to LM Evaluation Harness
Welcome and thank you for your interest in the LM Evaluation Harness! We welcome contributions and feedback and appreciate your time spent with our library, and hope you find it useful!
## Important Resources
There are several places information about LM Evaluation Harness is ... | {
"source": "simplescaling/s1",
"title": "eval/lm-evaluation-harness/docs/CONTRIBUTING.md",
"url": "https://github.com/simplescaling/s1/blob/main/eval/lm-evaluation-harness/docs/CONTRIBUTING.md",
"date": "2025-02-01T02:38:16",
"stars": 5696,
"description": "s1: Simple test-time scaling",
"file_size": 6071... |
# Eval Harness Documentation
Welcome to the docs for the LM Evaluation Harness!
## Table of Contents
* To learn about the public interface of the library, as well as how to evaluate via the command line or as integrated into an external library, see the [Interface](./interface.md).
* To learn how to add a new librar... | {
"source": "simplescaling/s1",
"title": "eval/lm-evaluation-harness/docs/README.md",
"url": "https://github.com/simplescaling/s1/blob/main/eval/lm-evaluation-harness/docs/README.md",
"date": "2025-02-01T02:38:16",
"stars": 5696,
"description": "s1: Simple test-time scaling",
"file_size": 858
} |
# Decontamination
## Usage
The provided directory should contain
the ngram files and info.json produced in "Pile Ngram Generation" further down.
```bash
python -m lm_eval \
--model gpt2 \
--device 0 \
--tasks sciq
```
## Background
Downstream evaluations test model generalization, and are less useful wh... | {
"source": "simplescaling/s1",
"title": "eval/lm-evaluation-harness/docs/decontamination.md",
"url": "https://github.com/simplescaling/s1/blob/main/eval/lm-evaluation-harness/docs/decontamination.md",
"date": "2025-02-01T02:38:16",
"stars": 5696,
"description": "s1: Simple test-time scaling",
"file_size"... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.