mirror of
https://github.com/Magnus167/msyrs.git
synced 2025-08-20 04:20:00 +00:00
working with docs
This commit is contained in:
parent
91aa5762d6
commit
6bb51905b8
8
Cargo.lock
generated
8
Cargo.lock
generated
@ -2559,18 +2559,18 @@ dependencies = [
|
||||
|
||||
[[package]]
|
||||
name = "serde"
|
||||
version = "1.0.214"
|
||||
version = "1.0.215"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "f55c3193aca71c12ad7890f1785d2b73e1b9f63a0bbc353c08ef26fe03fc56b5"
|
||||
checksum = "6513c1ad0b11a9376da888e3e0baa0077f1aed55c17f50e7b2397136129fb88f"
|
||||
dependencies = [
|
||||
"serde_derive",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "serde_derive"
|
||||
version = "1.0.214"
|
||||
version = "1.0.215"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "de523f781f095e28fa605cdce0f8307e451cc0fd14e2eb4cd2e98a355b147766"
|
||||
checksum = "ad1e866f866923f252f05c889987993144fb74e722403468a4ebd70c3cd756c0"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
|
@ -7,7 +7,7 @@ edition = "2021"
|
||||
reqwest = { version = "0.12.9", features = ["blocking", "json"] }
|
||||
serde_json = "1.0"
|
||||
serde_urlencoded = "0.7"
|
||||
serde = { version = "1.0", features = ["derive"] }
|
||||
serde = { version = "1.0.215", features = ["derive"] }
|
||||
polars = { version = "0.44.2", features = ["lazy"] }
|
||||
rand = "0.8"
|
||||
threadpool = "1.8.1"
|
||||
|
35
README.md
Normal file
35
README.md
Normal file
@ -0,0 +1,35 @@
|
||||
# msyrs
|
||||
|
||||
A Rust implementation of the [Macrosynergy Python Package](https://github.com/macrosynergy/macrosynergy).
|
||||
|
||||
## Status
|
||||
|
||||
- Docs
|
||||
- [x] Building
|
||||
- [ ] written?
|
||||
|
||||
- Download
|
||||
- [x] OAuth workflow
|
||||
- [x] Heartbeat
|
||||
- [x] Get Catalogue
|
||||
- [x] Get Generic DQ Time Series
|
||||
- [x] Get JPMaQS Indicators (formatted as array)
|
||||
- [x] Get JPMaQS Indicators as Polars DataFrame
|
||||
- [x] Add QDF option
|
||||
- [ ] Save to disk functionality
|
||||
|
||||
- Utils
|
||||
- [ ] Reduce DF
|
||||
- [ ] Reduce DF by ticker
|
||||
- [ ] Apply Blacklist
|
||||
- [ ] Get Blacklist
|
||||
- [ ] Apply Slip
|
||||
|
||||
|
||||
- Panel
|
||||
- [ ] Historic Volatility
|
||||
- [ ] Linear Composites
|
||||
- [ ] Make Relative Value
|
||||
- [ ] Imputers
|
||||
- [ ]
|
||||
|
6
docs/Cargo.toml
Normal file
6
docs/Cargo.toml
Normal file
@ -0,0 +1,6 @@
|
||||
[package]
|
||||
name = "docs"
|
||||
version = "0.1.0"
|
||||
edition = "2021"
|
||||
|
||||
[dependencies]
|
14
docs/src/lib.rs
Normal file
14
docs/src/lib.rs
Normal file
@ -0,0 +1,14 @@
|
||||
pub fn add(left: u64, right: u64) -> u64 {
|
||||
left + right
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use super::*;
|
||||
|
||||
#[test]
|
||||
fn it_works() {
|
||||
let result = add(2, 2);
|
||||
assert_eq!(result, 4);
|
||||
}
|
||||
}
|
@ -1,10 +1,8 @@
|
||||
use crate::download::oauth_client::OAuthClient;
|
||||
use crate::download::requester::DQRequester;
|
||||
use crate::download::timeseries::DQTimeSeries;
|
||||
use crate::download::timeseries::DQTimeSeriesResponse;
|
||||
use crate::download::timeseries::DQTimeseriesRequestArgs;
|
||||
use crate::download::timeseries::JPMaQSIndicator;
|
||||
use rayon::prelude::*;
|
||||
use std::error::Error;
|
||||
|
||||
const DEFAULT_JPMAQS_METRICS: [&str; 4] = ["value", "grading", "eop_lag", "mop_lag"];
|
||||
@ -114,53 +112,21 @@ impl JPMaQSDownload {
|
||||
|
||||
let expressions = construct_expressions(download_args.tickers, download_args.metrics);
|
||||
assert!(all_jpmaq_expressions(expressions.clone()));
|
||||
let dqts_vec = self.get_expressions(expressions)?;
|
||||
|
||||
// println!("Retrieved {} time series", -- sum[dqts_vec.iter().map(|dqts| dqts.len())]);
|
||||
println!("Retrieved all time series",);
|
||||
// say pausing for 30 seconds
|
||||
println!("Pausing for 10 seconds");
|
||||
std::thread::sleep(std::time::Duration::from_secs(10));
|
||||
println!("Resuming");
|
||||
let start = std::time::Instant::now();
|
||||
let dq_download_args = DQTimeseriesRequestArgs {
|
||||
expressions: expressions,
|
||||
start_date: download_args.start_date,
|
||||
end_date: download_args.end_date,
|
||||
..Default::default()
|
||||
};
|
||||
|
||||
let indicators = dq_response_vec_to_jpmaqs_indicators(dqts_vec);
|
||||
let result = self
|
||||
.requester
|
||||
.get_timeseries_as_jpmaqs_indicators(dq_download_args);
|
||||
|
||||
println!(
|
||||
"Converted time series to indicators in {:?}",
|
||||
start.elapsed()
|
||||
);
|
||||
println!("Pausing for 10 seconds");
|
||||
std::thread::sleep(std::time::Duration::from_secs(10));
|
||||
println!("Resuming");
|
||||
Ok(indicators)
|
||||
match result {
|
||||
Ok(indicators) => Ok(indicators),
|
||||
Err(e) => Err(e),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// fn dq_response_vec_to_jpmaqs_indicators(
|
||||
// dqts_vec: Vec<DQTimeSeriesResponse>,
|
||||
// ) -> Vec<JPMaQSIndicator> {
|
||||
// let mut indicators: Vec<JPMaQSIndicator> = Vec::new();
|
||||
// for dqts in dqts_vec {
|
||||
// indicators.extend(
|
||||
// dqts.consume_to_grouped_by_ticker() // moves the values to free up memory
|
||||
// .into_iter()
|
||||
// .filter_map(|tsv| JPMaQSIndicator::new(tsv).ok()),
|
||||
// );
|
||||
// }
|
||||
|
||||
// indicators
|
||||
// }
|
||||
fn dq_response_vec_to_jpmaqs_indicators(
|
||||
dqts_vec: Vec<DQTimeSeriesResponse>,
|
||||
) -> Vec<JPMaQSIndicator> {
|
||||
dqts_vec
|
||||
.into_par_iter()
|
||||
.flat_map(|dqts| {
|
||||
dqts.consume_to_grouped_by_ticker()
|
||||
.into_iter()
|
||||
.filter_map(|tsv| JPMaQSIndicator::new(tsv).ok())
|
||||
.collect::<Vec<JPMaQSIndicator>>()
|
||||
})
|
||||
.collect()
|
||||
}
|
||||
|
@ -146,6 +146,24 @@ impl DQRequester {
|
||||
&mut self,
|
||||
args: DQTimeseriesRequestArgs,
|
||||
) -> Result<Vec<DQTimeSeriesResponse>, Box<dyn Error>> {
|
||||
let max_retries = 5;
|
||||
println!(
|
||||
"Invoking ParallelRequester for {:?} expressions",
|
||||
args.expressions.len()
|
||||
);
|
||||
let mut pq = ParallelRequester::new(self.oauth_client.clone());
|
||||
let response_texts = match pq.request_expressions(args, max_retries) {
|
||||
Ok(r) => r,
|
||||
Err(e) => return Err(e),
|
||||
};
|
||||
let dqts_vec: Vec<DQTimeSeriesResponse> = parse_response_texts(response_texts);
|
||||
Ok(dqts_vec)
|
||||
}
|
||||
|
||||
pub fn get_timeseries_as_jpmaqs_indicators(
|
||||
&mut self,
|
||||
args: DQTimeseriesRequestArgs,
|
||||
) -> Result<Vec<JPMaQSIndicator>, Box<dyn Error>> {
|
||||
let max_retries = 5;
|
||||
println!(
|
||||
"Invoking ParallelRequester for {:?} expressions",
|
||||
@ -154,7 +172,7 @@ impl DQRequester {
|
||||
let mut pq = ParallelRequester::new(self.oauth_client.clone());
|
||||
let start = std::time::Instant::now();
|
||||
|
||||
let response_texts = match pq.request_expressions(args, max_retries) {
|
||||
let mut response_texts = match pq.request_expressions(args, max_retries) {
|
||||
Ok(r) => r,
|
||||
Err(e) => return Err(e),
|
||||
};
|
||||
@ -163,13 +181,19 @@ impl DQRequester {
|
||||
start.elapsed()
|
||||
);
|
||||
|
||||
let dqts_vec: Vec<DQTimeSeriesResponse> = parse_response_texts(response_texts);
|
||||
// sleep for 10 seconds
|
||||
println!("Pausing for 10 seconds");
|
||||
std::thread::sleep(std::time::Duration::from_secs(10));
|
||||
println!("Resuming - parsing response texts to JPMaQSIndicators");
|
||||
|
||||
let jpmaqs_indicators: Vec<JPMaQSIndicator> =
|
||||
parse_response_texts_to_jpmaqs_indicators(&mut response_texts);
|
||||
// Sleep for 10 seconds
|
||||
println!("Pausing for 10 seconds");
|
||||
std::thread::sleep(std::time::Duration::from_secs(10));
|
||||
println!("Resuming");
|
||||
|
||||
Ok(dqts_vec)
|
||||
Ok(jpmaqs_indicators)
|
||||
}
|
||||
}
|
||||
|
||||
@ -188,6 +212,54 @@ fn parse_response_texts(response_texts: Vec<String>) -> Vec<DQTimeSeriesResponse
|
||||
})
|
||||
.collect()
|
||||
}
|
||||
|
||||
fn parse_response_texts_to_jpmaqs_indicators(
|
||||
response_texts: &mut Vec<String>,
|
||||
) -> Vec<JPMaQSIndicator> {
|
||||
// Create an empty hashmap of jpmaqs indicators
|
||||
let mut jpmaqs_indicators_map: std::collections::HashMap<String, JPMaQSIndicator> =
|
||||
std::collections::HashMap::new();
|
||||
|
||||
// Iterate over the response texts by taking and consuming each element
|
||||
while let Some(response_text) = response_texts.pop() {
|
||||
// Attempt to deserialize the response text
|
||||
let json_res = serde_json::from_str::<DQTimeSeriesResponse>(&response_text);
|
||||
|
||||
// Free the memory occupied by response_text immediately after parsing
|
||||
drop(response_text);
|
||||
|
||||
let dqts_res = match json_res {
|
||||
Ok(dqts) => dqts,
|
||||
Err(err) => {
|
||||
eprintln!("Failed to deserialize response: {}", err);
|
||||
continue;
|
||||
}
|
||||
};
|
||||
|
||||
let grouped_by_ticker = dqts_res.consume_to_grouped_by_ticker();
|
||||
|
||||
for ts_vec in grouped_by_ticker {
|
||||
let curr_ticker = ts_vec[0].get_ticker().unwrap();
|
||||
if let Some(existing_jpmaqs_indicator) = jpmaqs_indicators_map.get_mut(&curr_ticker) {
|
||||
for tsv in ts_vec {
|
||||
match existing_jpmaqs_indicator.add_timeseries(tsv) {
|
||||
Ok(_) => {}
|
||||
Err(e) => {
|
||||
eprintln!("Failed to add timeseries: {}", e);
|
||||
}
|
||||
}
|
||||
}
|
||||
} else {
|
||||
let jpmaqs_indicator = JPMaQSIndicator::new(ts_vec);
|
||||
jpmaqs_indicators_map.insert(curr_ticker.into(), jpmaqs_indicator.unwrap());
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
println!("Number of responses left: {}", response_texts.len());
|
||||
jpmaqs_indicators_map.into_iter().map(|(_, v)| v).collect()
|
||||
}
|
||||
|
||||
#[allow(dead_code)]
|
||||
fn main() {
|
||||
let client_id = std::env::var("DQ_CLIENT_ID").unwrap();
|
||||
|
@ -153,6 +153,7 @@ impl Attribute {
|
||||
}
|
||||
|
||||
/// Get the metric from the expression
|
||||
#[allow(dead_code)]
|
||||
pub fn get_metric(&self) -> Result<String, Box<dyn Error>> {
|
||||
if !self.expression.starts_with("DB(JPMAQS,") {
|
||||
return Err("Expression does not start with 'DB(JPMAQS,'".into());
|
||||
@ -330,6 +331,14 @@ impl JPMaQSIndicator {
|
||||
})
|
||||
}
|
||||
|
||||
pub fn add_timeseries(&mut self, timeseries: DQTimeSeries) -> Result<(), Box<dyn Error>> {
|
||||
if self.ticker != timeseries.get_ticker()? {
|
||||
return Err("Timeseries does not belong to the same ticker".into());
|
||||
}
|
||||
add_timeseries_to_df(&mut self.df, timeseries)?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
pub fn as_qdf(&self) -> Result<DataFrame, Box<dyn Error>> {
|
||||
let mut qdf = self.df.clone();
|
||||
let (cid, xcat) = match self.ticker.split_once('_') {
|
||||
@ -405,6 +414,23 @@ fn timeseries_list_to_dataframe(
|
||||
Ok(output_df)
|
||||
}
|
||||
|
||||
fn add_timeseries_to_df(
|
||||
df: &mut DataFrame,
|
||||
timeseries: DQTimeSeries,
|
||||
) -> Result<(), Box<dyn Error>> {
|
||||
let mut new_df = timeseries.to_dataframe()?;
|
||||
let curr_metric = timeseries.get_metric()?;
|
||||
let column_name = new_df.get_column_names()[1].to_string();
|
||||
new_df
|
||||
.rename(&column_name, curr_metric.into())
|
||||
.expect("Failed to rename column");
|
||||
|
||||
*df = df
|
||||
.left_join(&new_df, ["real_date"], ["real_date"])
|
||||
.expect("Left join failed");
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn sort_qdf_columns(qdf: &mut DataFrame) -> Result<(), Box<dyn Error>> {
|
||||
let index_columns = ["real_date", "cid", "xcat"];
|
||||
let known_metrics = ["value", "grading", "eop_lag", "mop_lag"];
|
||||
|
Loading…
x
Reference in New Issue
Block a user