From 10489f71e5d0228e8028b5fa6dde7ce52a3dd982 Mon Sep 17 00:00:00 2001 From: Alejandro Velez Date: Wed, 8 Jan 2025 08:35:21 -0500 Subject: [PATCH 1/2] mend --- tdc/test/test_model_server.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tdc/test/test_model_server.py b/tdc/test/test_model_server.py index 3dcf1330..a9ac3ddb 100644 --- a/tdc/test/test_model_server.py +++ b/tdc/test/test_model_server.py @@ -108,7 +108,7 @@ def testscGPT(self): ) # Convert to numpy array tokenized_data = tokenizer.tokenize_cell_vectors( adata.X.toarray(), gene_ids) - first_embed = model(tokenized_data[0][1]).last_hidden_state + first_embed = model(tokenized_data[0]).last_hidden_state self.assertEqual(first_embed.shape[0], len(tokenized_data[0][0])) def testGeneformerTokenizer(self): From 23e916b4bdf2b19b37606c7aba774cc044052db0 Mon Sep 17 00:00:00 2001 From: Alejandro Velez Date: Wed, 8 Jan 2025 10:10:28 -0500 Subject: [PATCH 2/2] mend --- tdc/test/test_model_server.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tdc/test/test_model_server.py b/tdc/test/test_model_server.py index a9ac3ddb..3dcf1330 100644 --- a/tdc/test/test_model_server.py +++ b/tdc/test/test_model_server.py @@ -108,7 +108,7 @@ def testscGPT(self): ) # Convert to numpy array tokenized_data = tokenizer.tokenize_cell_vectors( adata.X.toarray(), gene_ids) - first_embed = model(tokenized_data[0]).last_hidden_state + first_embed = model(tokenized_data[0][1]).last_hidden_state self.assertEqual(first_embed.shape[0], len(tokenized_data[0][0])) def testGeneformerTokenizer(self):