1
- use std:: { collections :: HashMap , num:: ParseIntError } ;
1
+ use std:: num:: ParseIntError ;
2
2
3
3
use axum:: {
4
4
extract:: { Query , State } ,
@@ -28,22 +28,29 @@ struct IndexQuery {
28
28
29
29
#[ derive( Serialize ) ]
30
30
#[ serde( untagged) ]
31
- enum Counts {
32
- Normalized ( HashMap < String , f64 > ) ,
33
- Raw ( HashMap < String , i32 > ) ,
31
+ enum Values {
32
+ Normalized ( Vec < f64 > ) ,
33
+ Raw ( Vec < i32 > ) ,
34
34
}
35
35
36
36
#[ derive( Serialize ) ]
37
37
struct Run {
38
38
id : i32 ,
39
- counts : Counts ,
39
+ values : Values ,
40
40
}
41
41
42
42
#[ derive( Serialize ) ]
43
- struct IndexBody {
43
+ #[ serde( rename_all = "camelCase" ) ]
44
+ struct Counts {
45
+ feature_names : Vec < String > ,
44
46
runs : Vec < Run > ,
45
47
}
46
48
49
+ #[ derive( Serialize ) ]
50
+ struct IndexBody {
51
+ counts : Counts ,
52
+ }
53
+
47
54
pub fn router ( ) -> Router < Context > {
48
55
Router :: new ( ) . route ( "/counts" , get ( index) )
49
56
}
@@ -123,15 +130,9 @@ async fn index(
123
130
. into_iter ( )
124
131
. zip ( normalized_counts. axis_iter ( Axis ( 0 ) ) )
125
132
{
126
- let counts = feature_names
127
- . iter ( )
128
- . zip ( row)
129
- . map ( |( name, count) | ( name. clone ( ) , * count) )
130
- . collect ( ) ;
131
-
132
133
runs. push ( Run {
133
134
id,
134
- counts : Counts :: Normalized ( counts ) ,
135
+ values : Values :: Normalized ( row . to_vec ( ) ) ,
135
136
} ) ;
136
137
}
137
138
} else {
@@ -147,7 +148,7 @@ async fn index(
147
148
. map ( |( name, count) | ( name. clone ( ) , * count) )
148
149
. collect ( ) ;
149
150
150
- let normalized_counts = match normalization_method {
151
+ let normalized_counts_map = match normalization_method {
151
152
Normalize :: Fpkm => {
152
153
crate :: counts:: normalization:: fpkm:: calculate_fpkms ( & features, & counts)
153
154
. unwrap ( )
@@ -159,28 +160,32 @@ async fn index(
159
160
}
160
161
} ;
161
162
163
+ let normalized_counts = feature_names
164
+ . iter ( )
165
+ . map ( |name| normalized_counts_map[ name] )
166
+ . collect ( ) ;
167
+
162
168
runs. push ( Run {
163
169
id,
164
- counts : Counts :: Normalized ( normalized_counts) ,
170
+ values : Values :: Normalized ( normalized_counts) ,
165
171
} )
166
172
}
167
173
}
168
174
} else {
169
175
let chunks = counts. chunks_exact ( feature_names. len ( ) ) ;
170
176
171
177
for ( id, chunk) in run_ids. into_iter ( ) . zip ( chunks) {
172
- let counts = feature_names
173
- . iter ( )
174
- . zip ( chunk)
175
- . map ( |( name, count) | ( name. clone ( ) , * count) )
176
- . collect ( ) ;
177
-
178
178
runs. push ( Run {
179
179
id,
180
- counts : Counts :: Raw ( counts ) ,
180
+ values : Values :: Raw ( chunk . to_vec ( ) ) ,
181
181
} ) ;
182
182
}
183
183
}
184
184
185
- Ok ( Json ( IndexBody { runs } ) )
185
+ Ok ( Json ( IndexBody {
186
+ counts : Counts {
187
+ feature_names,
188
+ runs,
189
+ } ,
190
+ } ) )
186
191
}
0 commit comments