From cf3a7f89a96462259aa26fa3ee069dcc2051f9c9 Mon Sep 17 00:00:00 2001 From: raschedh <141400232+raschedh@users.noreply.github.com> Date: Mon, 24 Nov 2025 15:54:40 +0000 Subject: [PATCH 1/2] Update config.py (#312) --- helical/models/c2s/config.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/helical/models/c2s/config.py b/helical/models/c2s/config.py index 59b9bceb..cfdadef6 100644 --- a/helical/models/c2s/config.py +++ b/helical/models/c2s/config.py @@ -57,8 +57,8 @@ class Cell2SenConfig: If True, the attention implementation will be set to "flash_attention_2". If False, the attention implementation will be set to "sdpa". - max_genes: int = None - Maximum number of genes to use for the model. Default is None. + max_genes: int = 200 + Maximum number of genes to use for the model. Default is 200. If None, all genes will be used. If a number is provided, the genes will be sorted by expression level and the top max_genes will be used. @@ -73,7 +73,7 @@ def __init__( organism: str = None, perturbation_column: str = None, max_new_tokens: int = 200, - max_genes: int = None, + max_genes: int = 200, return_fit: bool = False, dtype: str = "bfloat16", model_size: str = "2B", @@ -138,4 +138,4 @@ def __init__( "use_flash_attn": use_flash_attn, "max_genes": max_genes, "device": device, - } \ No newline at end of file + } From 446fc985e2d11732e2e384b42d2600cc88045111 Mon Sep 17 00:00:00 2001 From: Benoit Putzeys <157973952+bputzeys@users.noreply.github.com> Date: Mon, 24 Nov 2025 16:55:40 +0100 Subject: [PATCH 2/2] Update pyproject.toml --- pyproject.toml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/pyproject.toml b/pyproject.toml index 808ca39e..3c4a737c 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -4,7 +4,7 @@ build-backend = "hatchling.build" [project] name = "helical" -version = "1.4.18" +version = "1.4.19" authors = [ { name="Helical Team", email="support@helical-ai.com" }, ]