Skip to content

Commit

Permalink
notebooks
Browse files Browse the repository at this point in the history
  • Loading branch information
mgdenno committed Nov 18, 2024
1 parent 4a753dc commit 6dd9571
Showing 1 changed file with 6 additions and 109 deletions.
115 changes: 6 additions & 109 deletions docs/sphinx/user_guide/notebooks/06_grouping_and_filtering.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -103,26 +103,13 @@
},
{
"cell_type": "code",
"execution_count": 1,
"execution_count": null,
"metadata": {
"tags": [
"hide-output"
]
},
"outputs": [
{
"name": "stderr",
"output_type": "stream",
"text": [
"24/11/17 10:28:01 WARN Utils: Your hostname, ubuntu3 resolves to a loopback address: 127.0.1.1; using 10.0.2.15 instead (on interface enp0s3)\n",
"24/11/17 10:28:01 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to another address\n",
"Setting default log level to \"WARN\".\n",
"To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).\n",
"24/11/17 10:28:01 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable\n",
"24/11/17 10:28:02 WARN Utils: Service 'SparkUI' could not bind on port 4040. Attempting port 4041.\n"
]
}
],
"outputs": [],
"source": [
"from pathlib import Path\n",
"import shutil\n",
Expand All @@ -139,113 +126,23 @@
},
{
"cell_type": "code",
"execution_count": 2,
"execution_count": null,
"metadata": {},
"outputs": [
{
"data": {
"text/html": [
"<div>\n",
"<style scoped>\n",
" .dataframe tbody tr th:only-of-type {\n",
" vertical-align: middle;\n",
" }\n",
"\n",
" .dataframe tbody tr th {\n",
" vertical-align: top;\n",
" }\n",
"\n",
" .dataframe thead th {\n",
" text-align: right;\n",
" }\n",
"</style>\n",
"<table border=\"1\" class=\"dataframe\">\n",
" <thead>\n",
" <tr style=\"text-align: right;\">\n",
" <th></th>\n",
" <th>name</th>\n",
" <th>description</th>\n",
" <th>url</th>\n",
" </tr>\n",
" </thead>\n",
" <tbody>\n",
" <tr>\n",
" <th>0</th>\n",
" <td>p0_2_location_example</td>\n",
" <td>Example evaluation datsets with 2 USGS gages</td>\n",
" <td>s3a://ciroh-rti-public-data/teehr-data-warehou...</td>\n",
" </tr>\n",
" <tr>\n",
" <th>1</th>\n",
" <td>p1_camels_daily_streamflow</td>\n",
" <td>Daily average streamflow at ther Camels basins</td>\n",
" <td>s3a://ciroh-rti-public-data/teehr-data-warehou...</td>\n",
" </tr>\n",
" <tr>\n",
" <th>2</th>\n",
" <td>p2_camels_hourly_streamflow</td>\n",
" <td>Hourly instantaneous streamflow at ther Camels...</td>\n",
" <td>s3a://ciroh-rti-public-data/teehr-data-warehou...</td>\n",
" </tr>\n",
" <tr>\n",
" <th>3</th>\n",
" <td>p3_usgs_hourly_streamflow</td>\n",
" <td>Hourly instantaneous streamflow at USGS CONUS ...</td>\n",
" <td>s3a://ciroh-rti-public-data/teehr-data-warehou...</td>\n",
" </tr>\n",
" </tbody>\n",
"</table>\n",
"</div>"
],
"text/plain": [
" name \\\n",
"0 p0_2_location_example \n",
"1 p1_camels_daily_streamflow \n",
"2 p2_camels_hourly_streamflow \n",
"3 p3_usgs_hourly_streamflow \n",
"\n",
" description \\\n",
"0 Example evaluation datsets with 2 USGS gages \n",
"1 Daily average streamflow at ther Camels basins \n",
"2 Hourly instantaneous streamflow at ther Camels... \n",
"3 Hourly instantaneous streamflow at USGS CONUS ... \n",
"\n",
" url \n",
"0 s3a://ciroh-rti-public-data/teehr-data-warehou... \n",
"1 s3a://ciroh-rti-public-data/teehr-data-warehou... \n",
"2 s3a://ciroh-rti-public-data/teehr-data-warehou... \n",
"3 s3a://ciroh-rti-public-data/teehr-data-warehou... "
]
},
"execution_count": 2,
"metadata": {},
"output_type": "execute_result"
}
],
"outputs": [],
"source": [
"# List the evaluations in the S3 bucket\n",
"ev.list_s3_evaluations()"
]
},
{
"cell_type": "code",
"execution_count": 3,
"execution_count": null,
"metadata": {
"tags": [
"hide-output"
]
},
"outputs": [
{
"name": "stderr",
"output_type": "stream",
"text": [
"24/11/17 10:28:09 WARN MetricsConfig: Cannot locate configuration: tried hadoop-metrics2-s3a-file-system.properties,hadoop-metrics2.properties\n",
"24/11/17 10:28:33 WARN SparkStringUtils: Truncated the string representation of a plan since it was too large. This behavior can be adjusted by setting 'spark.sql.debug.maxToStringFields'.\n",
" \r"
]
}
],
"outputs": [],
"source": [
"ev.clone_from_s3(\n",
" evaluation_name=\"p1_camels_daily_streamflow\",\n",
Expand Down

0 comments on commit 6dd9571

Please sign in to comment.