X-Git-Url: https://feistymeow.org/gitweb/?a=blobdiff_plain;f=scripts%2Fagenda%2Finfo_overload_report.sh;h=793f6566e1252921588a256aa6cb5179292f3590;hb=e6ce0aa508c3ad1beb167fdf20732b24f383c651;hp=a2f89349398941c57c4e804dc18cfac1cb0511a3;hpb=1f21a6ea5179051319665aa5c9226ea1563bd87c;p=feisty_meow.git diff --git a/scripts/agenda/info_overload_report.sh b/scripts/agenda/info_overload_report.sh index a2f89349..793f6566 100644 --- a/scripts/agenda/info_overload_report.sh +++ b/scripts/agenda/info_overload_report.sh @@ -1,7 +1,11 @@ +#!/bin/bash # these metrics are how bogged down we are in to-do type items. -REPORT_FILE="$HOME/cloud/fred_stats/overload_history.txt" +# logged historical file where we append our latest report. +REPORT_FILE="$CLOUD_BASE/stats/overload_history.txt" + +#hmmm: check path validity? # given a path, this will find how many items are under it, ignoring svn and git files, plus # other patterns we happen to notice are not useful. @@ -19,7 +23,7 @@ function calculate_count() function calculate_weight() { local dir="$1"; shift - local weight=$(find "$dir" -type f -exec echo \"{}\" ';' 2>/dev/null | grep -v "\.svn" | grep -v "\.git"| grep -v "\.basket" | grep -v "\.version" | grep -v "\.keep" | xargs ls -al | awk '{ print $5 }' | paste -sd+ | bc 2>/dev/null) + local weight=$(find "$dir" -type f -exec echo \"{}\" ';' 2>/dev/null | grep -v "\.svn" | grep -v "\.git"| grep -v "\.basket" | grep -v "\.version" | grep -v "\.keep" | xargs ls -al | awk '{ print $5 }' | paste -sd+ - | bc 2>/dev/null) if [ -z "$weight" ]; then echo 0; else echo "$weight"; fi } @@ -28,7 +32,7 @@ function calculate_weight() function calculate_complexity() { local dir="$1"; shift - local complexity=$(find "$dir" -type d | wc -l) + local complexity=$(find "$dir" -type d | wc -l | tr -d ' ') if [ -z "$complexity" ]; then echo 0; else echo "$complexity"; fi } @@ -108,42 +112,77 @@ full_report+="count\tcomplexity\tweight (kb)\tcategory\n\ ================================================================\n\ " +#### + +# this group is for hierarchies which we expect to have a stable name. + #hmmm: don't fail if the hierarchy doesn't exist. # high priority stuff would be called urgent. -analyze_hierarchy_and_report ~/cloud/urgent "high priority (aieeee!)" +analyze_hierarchy_and_report $CLOUD_BASE/aaa_priority "high priority (aieeee!)" # notes are individual files of tasks, usually, although some are combined. -analyze_hierarchy_and_report ~/cloud/grunty_notes "grunty notes (externalities)" +analyze_hierarchy_and_report $CLOUD_BASE/grunty* "grunty (external facing) notes" + +# web site development tasks. +analyze_hierarchy_and_report $CLOUD_BASE/webular "web design (ideas and tasks)" # feisty notes are about feisty meow(r) concerns ltd codebase development. -analyze_hierarchy_and_report ~/cloud/feisty_notes "feisty meow notes (mondo coding)" +analyze_hierarchy_and_report $CLOUD_BASE/feisty_notes "feisty meow notes (mondo coding)" + +# metaverse notes are about our ongoing simulator development and LSL scripting. +analyze_hierarchy_and_report $CLOUD_BASE/metaverse "metaverse in cyberspace design and scripting" # home notes are a new top-level category; used to be under the grunty. -analyze_hierarchy_and_report ~/cloud/branch_road "hearth and home notes (branch road)" +analyze_hierarchy_and_report $CLOUD_BASE/branch_road "hearth and home notes (branch road)" # and then count up the things that we think will be cleaned soon, but one thing we have learned # unsorted files haven't been categorized yet. -analyze_hierarchy_and_report ~/cloud/disordered "unsorted files" +analyze_hierarchy_and_report $CLOUD_BASE/disordered "disordered and maybe deranged files" # we now consider the backlog of things to read to be a relevant fact. this is going to hose # up our weight accounting considerably. -analyze_hierarchy_and_report ~/cloud/reading "reading list (for a quiet afternoon)" +analyze_hierarchy_and_report $CLOUD_BASE/reading "reading list (for a quiet afternoon)" + +# bluesky is our brainstorming and wunderthinking area for new things. +analyze_hierarchy_and_report $CLOUD_BASE/blue_sky "blue sky is the limit ideas" + +#### + +# this group is found based on matching a pattern, since we may not use a single +# standardized name for it, or maybe there are prefixes for similar items (like project-). # scan all the items declared as active projects. -analyze_by_dir_patterns "active issues" ~/cloud/active* +analyze_by_dir_patterns "active issues" $CLOUD_BASE/*active* + +# vocation is a prefix for anything i consider part of my life's work. +analyze_by_dir_patterns "life's work and other oddities" $CLOUD_BASE/vocation* -# scan across all appropriately named project or research folders that live in the "cloud". -analyze_by_dir_patterns "running projects" ~/cloud/project* ~/cloud/research* +# scan across all appropriately named project or research folders. +analyze_by_dir_patterns "running projects" $CLOUD_BASE/*project* $CLOUD_BASE/*research* # look for our mad scientist style efforts. -analyze_by_dir_patterns "lab experiments" ~/cloud/experiment* +analyze_by_dir_patterns "lab experiments" $CLOUD_BASE/*experiment* + +# we also are running some businesses. get those notes in. +analyze_by_dir_patterns "business ops" $CLOUD_BASE/*business* + +# find any charity actions that need to be taken. +analyze_by_dir_patterns "charitable" $CLOUD_BASE/*charit* + +# rub alongside all the travel notes to see if any have interesting burrs. +analyze_by_dir_patterns "travel plans" $CLOUD_BASE/*travel* # snag any work related items for that category. -analyze_by_dir_patterns "jobby work tasks" ~/cloud/job* +analyze_by_dir_patterns "jobby work tasks" $CLOUD_BASE/job* # scan all the trivial project folders. -analyze_by_dir_patterns "trivialities" ~/cloud/trivia* +analyze_by_dir_patterns "trivialities and back burner items" $CLOUD_BASE/trivia* $CLOUD_BASE/backburn* + +# okay, fudge. if there are game tasks, then count them too. what are they, nothing? not so. +analyze_by_dir_patterns "play time and games" $CLOUD_BASE/*gaming* $CLOUD_BASE/*game* + +#### full_report+="================================================================\n\ "