This is a discussion about rational project planning. The major theme is that people should consider what their project premises are. What claims are they betting their project success on the correctness of? And why? This matter requires investigation and consideration, not just ignoring it.
By project I mean merely a goal-directed activity. It can be, but doesn't have to be, a business project or multi-person project. My primary focus is on larger projects, e.g. projects that take more than one day to finish.
The first part is discussion context. You may want to skip to the second part where I write an article/monologue with no one else talking. It explains a lot of important stuff IMO.
Gavin Palmer:
The most important problem is The Human Resource Problem. All other problems depend on the human resource problem. The Human Resource Problem consists of a set of smaller problems that are related. An important problem within that set is the communication problem: an inability to communicate. I classify that problem as a problem related to information technology and/or process. If people can obtain and maintain a state of mind which allows communication, then there are other problems within that set related to problems faced by any organization. Every organization is faced with problems related to hiring, firing, promotion, and demotion.
So every person encounters this problem. It is a universal problem. It will exist so long as there are humans. We each have the opportunity to recognize and remember this important problem in order to discover and implement processes and tools which can facilitate our ability to solve every problem which is solvable.
curi:
you haven't explained what the human resource problem is, like what things go in that category
Gavin Palmer:
The thought I originally had long ago - was that there are people willing and able to solve our big problems. We just don't have a sufficient mechanism for finding and organizing those people. But I have discovered that this general problem is related to ideas within any organization. The general problem is related to ideas within a company, a government, and even those encountered by each individual mind. The task of recruiting, hiring, firing, promoting, and demoting ideas can occur on multiple levels.
curi:
so you mean it like HR in companies? that strikes me as a much more minor problem than how rationality works.
Gavin Palmer:
If you want to end world hunger it's an HR problem.
curi:
it's many things including a rationality problem
curi:
and a free trade problem and a governance problem and a peace problem
curi:
all of which require rationality, which is why rationality is central
Gavin Palmer:
How much time have you actually put into trying to understand world hunger and the ways it could end?
Gavin Palmer:
How much time have you actually put into building anything? What's your best accomplishment as a human being?
curi:
are you mad?
GISTE:
so to summarize the discussion that Gavin started. Gavin described what he sees as the most important problem (the HR problem), where all other problems depend on it. curi disagreed by saying that how rationality works is a more important problem than the HR problem, and he gave reasons for it. Gavin disagreed by saying that for the goal of ending world hunger, the most important problem is the HR problem -- and he did not address curi's reasons. curi disagreed by saying that the goal of ending world hunger is many problems, all of which require rationality, making rationality the most important problem. Then Gavin asked curi about how much time he has spent on the world hunger problem and asked if he built anything and what his best accomplishments are. Gavin's response does not seem to connect to any of the previous discussion, as far as I can tell. So it's offtopic to the topic of what is the most important problem for the goal of ending world hunger. Maybe Gavin thinks it is on topic, but he didn't say why he thinks so. I guess that curi also noticed the offtopic thing, and that he guessed that Gavin is mad. then curi asked Gavin "are you mad?" as a way to try to address a bottleneck to this discussion. @Gavin Palmer is this how you view how the discussion went or do you have some differences from my view? if there are differences, then we could talk about those, which would serve to help us all get on the same page. And then that would help serve the purpose of reaching mutual understanding and agreement regarding whether or not the HR problem is the most important problem on which all other problems depend.
GISTE:
btw i think Gavin's topic is important. as i see it, it's goal is to figure out the relationships between various problems, to figure out which is the most important. i think that's important because it would serve the purpose of helping one figure out which problems to prioritize.
Gavin Palmer:
Here is a google doc linked to a 1-on-1 I had with GISTE (he gave me permission to share). I did get a little angry and was anxious about returning here today. I'm glad to see @curi did not get offended by my questions and asked a question. I am seeing the response after I had the conversation with GISTE. Thank you for your time.
https://docs.google.com/document/d/1XEztqEHLBAJ39HQlueKX3L4rVEGiZ4GEfBJUyXEgVNA/edit?usp=sharing
GISTE:
to be clear, regarding the 1 on 1 discussion linked above, whatever i said about curi are my interpretations. don't treat me as an authority on what curi thinks.
GISTE:
also, don't judge curi by my ideas/actions. that would be unfair to him. (also unfair to me)
JustinCEO:
Curi's response tells me he does not know how to solve world hunger.
JustinCEO:
Unclear to me how that judgment was arrived at
JustinCEO:
I'm reading
JustinCEO:
Lowercase c for curi btw
JustinCEO:
But I have thought about government, free trade, and peace very much. These aren't a root problem related to world hunger.
JustinCEO:
curi actually brought those up as examples of things that require rationality
JustinCEO:
And said that rationality was central
JustinCEO:
But you don't mention rationality in your statement of disagreement
JustinCEO:
You mention the examples but not the unifying theme
JustinCEO:
GISTE:
curi did not say those are root problems.
JustinCEO:
Ya 🙂
JustinCEO:
Ya GISTE got this point
JustinCEO:
I'm on phone so I'm pasting less than I might otherwise
JustinCEO:
another way to think about the world hunger problem is this: what are the bottlenecks to solving it? first name them, before trying to figure out which one is like the most systemic one.
JustinCEO:
I think the problem itself could benefit from a clear statement
GISTE:
That clear statement would include causes of (world) hunger. Right ? @JustinCEO
JustinCEO:
I mean a detailed statement would get into that issue some GISTE cuz like
JustinCEO:
You'd need to figure out what counts and what doesn't as an example of world hunger
JustinCEO:
What is in the class of world hunger and what is outside of it
JustinCEO:
And that involves getting into specific causes
JustinCEO:
Like presumably "I live in a first world country and have 20k in the bank but forgot to buy groceries this week and am hungry now" is excluded from most people's definitions of world hunger
JustinCEO:
I think hunger is basically a solved problem in western liberal capitalist democracies
JustinCEO:
People fake the truth of this by making up concepts called "food insecurity" that involve criteria like "occasionally worries about paying for groceries" and calling that part of a hunger issue
JustinCEO:
Thinking about it quickly, I kinda doubt there is a "world hunger" problem per se
GISTE:
yeah before you replied to my last comment, i immediately thought of people who choose to be hungry, like anorexic people. and i think people who talk about world hunger are not including those situations.
JustinCEO:
There's totally a Venezuela hunger problem or a Zimbabwe hunger problem tho
JustinCEO:
But not really an Ohio or Kansas hunger problem
JustinCEO:
Gavin
I try to be pragmatic. If your solution depends on people being rational, then the solution probably will not work. My solution does depend on rational people, but the number of rational people needed is very small
GISTE:
There was one last comment by me that did not get included in the one on one discussion. Here it is. “so, you only want people on your team that already did a bunch of work to solve world hunger? i thought you wanted rational people, not necessarily people that already did a bunch of work to solve world hunger.”
JustinCEO:
What you think being rational is and what it involves could probably benefit from some clarification.
Anyways I think society mostly works to the extent people are somewhat rational in a given context.
JustinCEO:
I regard violent crime for the purpose of stealing property as irrational
JustinCEO:
For example
JustinCEO:
Most people agree
JustinCEO:
So I can form a plan to walk down my block with my iPhone and not get robbed, and this plan largely depends on the rationality of other people
JustinCEO:
Not everyone agrees with my perspective
JustinCEO:
The cop car from the local precinct that is generally parked at the corner is also part of my plan
JustinCEO:
But my plan largely depends on the rationality of other people
JustinCEO:
If 10% or even 5% of people had a pro property crime perspective, the police could not really handle that and I would have to change my plans
Gavin Palmer:
World hunger is just an example of a big problem which depends on information technology related to the human resource problem. My hope is that people interested in any big problem could come to realize that information technology related to the human resource problem is part of the solution to the big problem they are interested in as well as other big problems.
Gavin Palmer:
So maybe "rationality" is related to what I call "information technology".
JustinCEO:
the rationality requirements of my walking outside with phone plan are modest. i can't plan to e.g. live in a society i would consider more moral and just (where e.g. a big chunk of my earnings aren't confiscated and wasted) cuz there's not enough people in the world who agree with me on the relevant issues to facilitate such a plan.
JustinCEO:
anyways regarding specifically this statement
JustinCEO:
If your solution depends on people being rational, then the solution probably will not work.
JustinCEO:
i wonder if the meaning is If your solution depends on [everyone] being [completely] rational, then the solution probably will not work.
Gavin Palmer:
There is definitely some number/percentage I have thought about... like I only need 10% of the population to be "rational".
GISTE:
@Gavin Palmer can you explain your point more? what i have in mind doens't seem to match your statement. so like if 90% of the people around me weren't rational (like to what degree exactly?), then they'd be stealing and murdering so much that the police couldn't stop them.
JustinCEO:
@Gavin Palmer based on the stuff you said so far and in the google doc regarding wanting to work on important problems, you may appreciate this post
JustinCEO:
https://curi.us/2029-the-worlds-biggest-problems
JustinCEO:
Gavin says
A thing that is sacred is deemed worthy of worship. And worship is based in the words worth and ship. And so a sacred word is believed to carry great worth in the mind of the believer. So I can solve world hunger with the help of people who are able and willing. Solving world hunger is not an act done by people who uphold the word rationality above all other words.
JustinCEO:
the word doesn't matter but the concept surely does for problem-solving effectiveness
JustinCEO:
people who don't value rationality can't solve much of anything
nikluk:
Re rationality. Have you read this article and do you agree with what it says, @Gavin Palmer ?
https://fallibleideas.com/reason
GISTE:
So maybe "rationality" is related to what I call "information technology".
can you say more about that relationship? i'm not sure what you have in mind. i could guess but i think it'd be a wild guess that i'm not confident would be right. (so like i could steelman your position but i could easily be adding in my own ideas and ruin it. so i'd rather avoid that.) @Gavin Palmer
Gavin Palmer:
so like if 90% of the people around me weren't rational (like to what degree exactly?), then they'd be stealing and murdering so much that the police couldn't stop them.
I think the image of the elephant rider portrayed by Jonathan Haidt is closer to the truth when it comes to some word like rationality and reason. I actually value something like compassion above a person's intellect: and I really like people who have both. There are plenty of idiots in the world who are not going to try and steal from you or murder you. I'm just going to go through these one by one when able.
Gavin Palmer:
https://curi.us/2029-the-worlds-biggest-problems
Learning to think is very important. There were a few mistakes in that article. The big one in my opinion is the idea that 2/3 of the people can change things. On the contrary our government systems do not have any mechanism in place to learn what 2/3 of the people actually want nor any ability to allow the greatest problem solvers to influence those 2/3 of the people. We aren't even able to recognize the greatest problem solvers. Another important problem is technology which allows for this kind of information sharing so that we can actually know what the people think and we can allow the greatest problem solvers to be heard. We want that signal to rise above the noise.
The ability to solve problems is like a muscle. For me - reading books does not help me build that muscle - they only help me find better words for describing the strategies and processes which I have developed through trial and error. I am not the smartest person - I learn from trial and error.
curi:
To answer the questions: I have thought about many big problems, such as aging death, AGI, and coercive parenting/education. Yes I've considered world hunger too, though not as a major focus. I'm an (experienced) intellectual. My accomplishments are primarily in philosophy research re issues like how learning and rational discussion work. I do a lot of educational writing and discussion. https://elliottemple.com
curi:
You're underestimating the level of outlier you're dealing with here, and jumping to conclusions too much.
Gavin Palmer:
https://fallibleideas.com/reason
It's pretty good. But science without engineering is dead. That previous sentence reminds me of "faith without works is dead". I'm not a huge fan of science for the sake of science. I'm a fan of engineering and the science that helps us do engineering.
curi:
i don't thikn i have anything against engineering.
Gavin Palmer:
I'm just really interested in finding people who want to help do the engineering. It's my bias. Even more - it's my passion and my obsession.
Gavin Palmer:
Thinking and having conversations is fun though.
Gavin Palmer:
But sometimes it can feel aimless if I'm not building something useful.
curi:
My understanding of the world, in big picture, is that a large portion of all efforts at engineering and other getting-stuff-done type work are misdirected and useless or destructive.
curi:
This is for big hard problems. The productiveness of practical effort is higher for little things like making dinner today.
curi:
The problem is largely not the engineering itself but the ideas guiding it – the goals and plan.
Gavin Palmer:
I worked for the Army's missile defense program for 6 years when I graduated from college. I left because of the reason you point out. My hope was that I would be able to change things from within.
curi:
So for example in the US you may agree with me that at least around half of political activism is misdirected to goals with low or negative value. (either the red tribe or blue tribe work is wrong, plus some of the other work too)
Gavin Palmer:
Even the ones I agree with and have volunteered for are doing a shit job.
curi:
yeah
curi:
i have found a decent number of people want to "change the world" or make some big improvement, but they can't agree amongst themselves about what changes to make, and some of them are working against others. i think sorting that mess out, and being really confident the projects one works on are actually good, needs to come before implementation.
curi:
i find most people are way too eager to jump into their favored cause without adequately considering why people disagree with it and sorting out all the arguments for all sides.
Gavin Palmer:
There are many tools that don't exist which could exist. And those tools could empower any organization and their goal(s).
curi:
no doubt.
curi:
software is pretty new and undeveloped. adequate tools are much harder to name than inadequate ones.
Gavin Palmer:
adequate tools are much harder to name than inadequate ones.
I don't know what that means.
curi:
we could have much better software tools for ~everything
curi:
"~" means "approximately"
JustinCEO:
Twitter can't handle displaying tweets well. MailMate performance gets sluggish with too many emails. Most PDF software can't handle super huge PDFs well. Workout apps can't use LIDAR to tell ppl if their form is on point
curi:
Discord is clearly a regression from IRC in major ways.
Gavin Palmer:
🤦♂️
JustinCEO:
?
JustinCEO:
i find your face palm very unclear @Gavin Palmer; hope you elaborate!
Gavin Palmer:
I find sarcasm very unclear. That's the only way I know how to interpret the comments about Twitter, MailMate, PDF, LIDAR, Discord, IRC, etc.
curi:
I wasn't being sarcastic and I'm confident Justin also meant what he said literally and seriously.
Gavin Palmer:
Ok - thanks for the clarification.
JustinCEO:
ya my statements were made earnestly
JustinCEO:
re: twitter example
JustinCEO:
twitter makes it harder to have a decent conversation cuz it's not good at doing conversation threading
JustinCEO:
if it was better at this, maybe people could keep track of discussions better and reach agreement more easily
Gavin Palmer:
Well - I have opinions about Twitter. But to be honest - I am also trying to look at what this guy is doing:
https://github.com/erezsh/portal-radar
It isn't a good name in my opinion - but the idea is related to having some bot collect discord data so that there can be tools which help people find the signal in the noise.
curi:
are you aware of http://bash.org ? i'm serious about major regressions.
JustinCEO:
i made an autologging system to make discord chat logs on this server so people could pull information (discussions) out of them more easily
JustinCEO:
but alas it's a rube goldberg machine of different tools running together in a VM, not something i can distribute
Gavin Palmer:
Well - it's a good goal. I'm looking to add some new endpoints in a pull request to the github repo I linked above. Then I could add some visualizations.
Another person has built a graphql backend (which he isn't sharing open source) and I have created some of my first react/d3 components to visualize his data.
https://portal-projects.github.io/users/
Gavin Palmer:
I think you definitely want to write the code in a way that it can facilitate collaboration.
curi:
i don't think this stuff will make much difference when people don't know what a rational discussion is and don't want one.
curi:
and don't want to use tools that already exist like google groups.
curi:
which is dramatically better than twitter for discussion
Gavin Palmer:
I'm personally interested in something which I have titled "Personality Targeting with Machine Learning".
Gavin Palmer:
My goal isn't to teach people to be rational - it is to try and find people who are trying to be rational.
curi:
have you identified which philosophical schools of thought it's compatible and incompatible with? and therefore which you're betting on being wrong?
curi:
it = "Personality Targeting with Machine Learning".
Gavin Palmer:
Ideally it isn't hard coded or anything. I could create multiple personality profiles. Three of the markets I have thought about using the technology in would be online dating, recruiting, and security/defense.
curi:
so no?
Gavin Palmer:
If I'm understanding you - a person using the software could create a personality that mimics a historical person for example - and then parse social media in search of people who are saying similar things.
Gavin Palmer:
But I'm not exactly sure what point you are trying to make.
curi:
You are making major bets while being unaware of what they are. You may be wrong and wasting your time and effort, or even being doing something counterproductive. And you aren't very interested in this.
Gavin Palmer:
Well - from my perspective - I am not making any major bets. What is the worst case scenario?
curi:
An example worst case scenario would be that you develop an AGI by accident and it turns us all into paperclips.
Gavin Palmer:
I work with a very intelligent person that would laugh at that idea.
curi:
That sounds like an admission you're betting against it.
curi:
You asked for an example seemingly because you were unaware of any. You should be documenting what bets you're making and why.
Gavin Palmer:
I won't be making software that turns us all into paperclips.
curi:
Have you studied AI alignment?
Gavin Palmer:
I have been writing software for over a decade. I have been using machine learning for many months now. And I have a pretty good idea of how the technology I am using actually works.
curi:
So no?
Gavin Palmer:
No. But if it is crap - do you want to learn why it is crap?
curi:
I would if I agreed with it, though I don't. But a lot of smart people believe it.
curi:
They have some fairly sophisticated reasons, which I don't think it's reasonable to bet against from a position of ignorance.
Gavin Palmer:
Our ability to gauge if someone has understanding on a given subject is relative to how much understanding we have on that subject.
curi:
Roughly, sure. What's your point?
Gavin Palmer:
First off - I'm not sure AGI is even possible. I love to play with the idea. And I would love to get to a point where I get to help build a god. But I am not even close to doing that at this point in my career.
curi:
So what?
Gavin Palmer:
You think there is a risk I would build something that turns humans into paperclips.
curi:
I didn't say that.
Gavin Palmer:
You said that is the worst case scenario.
curi:
Yes. It's something you're betting against, apparently without much familiarity with the matter.
curi:
Given that you don't know much about it, you aren't in a reasonable position to judge how big a risk it is.
curi:
So I think you're making a mistake.
curi:
The bigger picture mistake is not trying to figure out what bets you're making and why.
curi:
Most projects have this flaw.
Gavin Palmer:
My software uses algorithms to classify input data.
curi:
So then, usually, somewhere on the list of thousands of bets being made, are a few bad ones.
curi:
Does this concept make sense to you?
Gavin Palmer:
Love is most important in my hierarchy of values.
Gavin Palmer:
If I used the word in a sentence I would still want to capitalize it.
curi:
is that intended to be an answer?
Gavin Palmer:
Yes - I treat Love in a magical way. And you don't like magical thinking. And so we have very different world views. They might even be incompatible. The difference between us is that I won't be paralyzed by my fears. And I will definitely make mistakes. But I will make more mistakes than you. The quality and quantity of my learning will be very different than yours. But I will also be reaping the benefits of developing new relationships with engineers, learning new technology/process, and building up my portfolio of open source software.
curi:
You accuse me of being paralyzed by fears. You have no evidence and don't understand me.
curi:
Your message is not loving or charitable.
curi:
You're heavily personalizing while knowing almost nothing about me.
JustinCEO:
i agree
JustinCEO:
also, magical thinking can't achieve anything
curi:
But I will also be reaping the benefits of developing new relationships with engineers
curi:
right now you seem to be trying to burn a bridge with an engineer.
curi:
you feel attacked in some way. you're experiencing some sort of conflict. do you want to use a rational problem solving method to try to address this?
curi:
J, taking my side here will result in him feeling ganged up on. I think it will be counterproductive psychologically.
doubtingthomas:
J, taking my side here will result in him feeling ganged up on. I think it will be counterproductive psychologically.
Good observation. Are you going to start taking these considerations into account in future conversations?
curi:
I knew that years ago. I already did take it into account.
curi:
please take this tangent to #fi
GISTE:
also, magical thinking can't achieve anything
@JustinCEO besides temporary nice feelings. Long term its bad though.
doubtingthomas:
yeah sure
JustinCEO:
ya sure GISTE, i meant achieve something in reality
curi:
please stop talking here. everyone but gavin
Gavin Palmer:
You talked about schools of philosophy, AI alignment, and identifying the hidden bets. That's a lot to request of someone.
curi:
Thinking about your controversial premises and civilizational risks, in some way instead of ignoring the matter, is too big an ask to expect of people before they go ahead with projects?
curi:
Is that what you mean?
Gavin Palmer:
I don't see how my premises are controversial or risky.
curi:
Slow down. Is that what you meant? Did I understand you?
Gavin Palmer:
I am OK with people thinking about premises and risks of an idea and discussing those. But in order to have that kind of discussion you would need to understand the idea. And in order to understand the idea - you have to ask questions.
curi:
it's hard to talk with you because of your repeated unwillingness to give direct answers or responses.
curi:
i don't know how to have a productive discussion under these conditions.
Gavin Palmer:
I will try to do better.
curi:
ok. can we back up?
Thinking about your controversial premises and civilizational risks, in some way instead of ignoring the matter, is too big an ask to expect of people before they go ahead with projects?
did i understand you, yes or no?
Gavin Palmer:
no
curi:
ok. which part(s) is incorrect?
Gavin Palmer:
The words controversial and civilizational are not conducive to communication.
curi:
why?
Gavin Palmer:
They indicate that you think you understand the premises and the risks and I don't know that you understand the idea I am trying to communicate.
curi:
They are just adjectives. They don't say what I understand about your project.
Gavin Palmer:
Why did you use them?
curi:
Because you should especially think about controversial premises rather than all premises, and civilizational risks more than all risks.
curi:
And those are the types of things that were under discussion.
curi:
A generic, unqualified term like "premises" or "risks" would not accurately represent the list of 3 examples "schools of philosophy, AI alignment, and identifying the hidden bets"
Gavin Palmer:
I don't see how schools of philosophy, AI alignment, and hidden bets are relevant. Those are just meaningless words in my mind. The meaning of those words in your mind may contain relevant points. And I would be willing to discuss those points as they relate to the project. But (I think) that would also require that you have some idea of what the software does and how it is done. To bring up these things before you understand the software seems very premature.
curi:
the details of your project are not relevant when i'm bringing up extremely generic issues.
curi:
e.g. there is realism vs idealism. your project takes one side, the other, or is compatible with both. i don't need to know more about your project to say this.
curi:
(or disagrees with both, though that'd be unusual)
curi:
it's similar with skepticism or not.
curi:
and moral relativism.
curi:
and strong empiricism.
curi:
one could go on. at length. and add a lot more using details of your project, too.
curi:
so, there exists some big list. it has stuff on it.
curi:
so, my point is that you ought to have some way of considering and dealing with this list.
curi:
some way of considering what's on it, figuring out which merit attention and how to prioritize that attention, etc.
curi:
you need some sort of policy, some way to think about it that you regard as adequate.
curi:
this is true of all projects.
curi:
this is one of the issues which has logical priority over the specifics of your project.
curi:
there are generic concepts about how to approach a project which take precedence over jumping into the details.
curi:
do you think you understand what i'm saying?
Gavin Palmer:
I think I understand this statement:
there are generic concepts about how to approach a project which take precedence over jumping into the details.
curi:
ok. do you agree with that?
Gavin Palmer:
I usually jump into the details. I'm not saying you are wrong though.
curi:
ok. i think looking at least a little at the big picture is really important, and that most projects lose a lot of effectiveness (or worse) due to failing to do this plus some common errors.
curi:
and not having any conscious policy at all regarding this issue (how to think about the many premises you are building on which may be wrong) is one of the common errors.
curi:
i think being willing to think about things like this is one of the requirements for someone who wants to be effective at saving/changing/helping the world (or themselves individually)
Gavin Palmer:
But I have looked at a lot of big picture things in my life.
curi:
cool. doesn't mean you covered all the key ones. but maybe it'll give you a head start on the project planning stuff.
Gavin Palmer:
So do you have an example of a project where it was done in a way that is satisfactory in your mind?
curi:
hmm. project planning steps are broadly unpublished and unavailable for the vast majority of projects. i think the short answer is no one is doing this right. this aspect of rationality is ~novel.
curi:
some ppl do a more reasonable job but it's really hard to tell what most ppl did.
curi:
u can look at project success as a proxy but i don't think that'll be informative in the way you want.
Gavin Palmer:
I'm going to break soon, but I would encourage you to think about some action items for you and I based around this ideal form of project planning. I have real-world experience with various forms of project planning to some degree or another.
curi's Monologue
curi:
the standard way to start is to brainstorm things on the list
curi:
after you get a bunch, you try to organize them into categories
curi:
you also consider what is a reasonable level of overhead for this, e.g. 10% of total project resource budget.
curi:
but a flat percentage is problematic b/c a lot of the work is general education stuff that is reusable for most projects. if you count your whole education, overhead will generally be larger than the project. if you only count stuff specific to this project, you can have a really small overhead and do well.
curi:
stuff like reading and understanding/remembering/taking-notes-on/etc one overview book of philosophy ideas is something that IMO should be part of being an educated person who has appropriate background knowledge. but many ppl haven't done it. if you assign the whole cost of that to a one project it can make the overhead ratio look bad.
curi:
unfortunately i think a lot of what's in that book would be wrong and ignore some more important but less famous ideas. but at least that'd be a reasonable try. most ppl don't even get that far.
curi:
certainly a decent number of ppl have done that. but i think few have ever consciously considered "which philosophy schools of thought does my project contradict? which am i assuming as premises and betting my project success on? and is that a good idea? do any merit more investigation before i make such a bet?" ppl have certainly considered such things in a disorganized, haphazard way, which sometimes manages to work out ok. idk that ppl have done this by design in that way i'm recommending.
curi:
this kind of analysis has large practical consequences, e.g. > 50% of "scientific research" is in contradiction to Critical Rationalist epistemology, which is one of the more famous philosophies of science.
curi:
IMO, consequently it doesn't work and the majority of scientists basically waste their careers.
curi:
most do it without consciously realizing they are betting their careers on Karl Popper being wrong.
curi:
many of them do it without reading any Popper book or being able to name any article criticizing Popper that they think is correct.
curi:
that's a poor bet to make.
curi:
even if Popper is wrong, one should have more information before betting against him like that.
curi:
another thing with scientists is the majority bet their careers on a claim along the lines of "college educations and academia are good"
curi:
this is a belief that some of the best scientists have disagreed with
curi:
a lot of them also have government funding underlying their projects and careers without doing a rational investigation of whether that may be a really bad, risky thing.
curi:
separate issue: broadly, most large projects try to use reason. part of the project is that problems come up and people try to do rational problem solving – use reason to solve the problems as they come up. they don't expect to predict and plan for every issue they're gonna face. there are open controversies about what reason is, how to use it, what problem solving methods are effective or ineffective, etc.
curi:
what the typical project does is go by common sense and intuition. they are basically betting the project on whatever concept of reason they picked up here and there from their culture being adequate. i regard this as a very risky bet.
curi:
and different project members have different conceptions of reason, and they are also betting on those being similar enough things don't fall apart.
curi:
commonly without even attempting to talk about the matter or put their ideas into words.
curi:
what happens a lot when people have unverbalized philosophy they picked up from their culture at some unknown time in the past is ... BIAS. they don't actually stick to any consistent set of ideas about reason. they change it around situationally according to their biases. that's a problem on top of some of the ideas floating around our culture being wrong (which is well known – everyone knows that lots of ppl's attempts at rational problem solving don't work well)
curi:
one of the problems in the field of reason is: when and how do you rationally end (or refuse to start) conversations without agreement. sometimes you and the other guy agree. but sometimes you don't, and the guy is saying "you're wrong and it's a big deal, so you shouldn't just shut your mind and refuse to consider more" and you don't want to deal with that endlessly but you also don't want to just be biased and stay wrong, so how do you make an objective decision? preferably is there something you could say that the other guy could accept as reasonable? (not with 100% success rate, some people gonna yell at you no matter what, but something that would convince 99% of people who our society considers pretty smart or reasonable?)
curi:
this has received very little consideration from anyone and has resulted in countless disputes when people disagree about whether it's appropriate to stop a discussion without giving further answers or arguments.
curi:
lots of projects have lots of strife over this specific thing.
curi:
i also was serious about AI risk being worth considering (for basically anything in the ballpark of machine learning, like classifying big data sets) even though i actually disagree with that one. i did consider it and think it merits consideration.
curi:
i think it's very similar to physicists in 1940 were irresponsible if they were doing work anywhere in the ballpark of nuclear stuff and didn't think about potential weapons.
curi:
another example of a project management issue is how does one manage a schedule? how full should a schedule be packed with activities? i think the standard common sense ways ppl deal with this are wrong and do a lot of harm (the basic error is overfilling schedules in a way which fails to account for variance in task completion times, as explained by Eliyahu Goldratt)
curi:
i meant there an individual person's schedule
curi:
similarly there is problem of organizing the entire project schedule and coordinating people and things. this has received a ton of attention from specialists, but i think most ppl have an attitude like "trust a standard view i learned in my MBA course. don't investigate rival viewpoints". risky.
curi:
a lot of other ppl have no formal education about the matter and mostly ... don't look it up and wing it.
curi:
even riskier!
curi:
i think most projects managers couldn't speak very intelligently about early start vs. late start for dependencies off the critical path.
curi:
and don't know that Goldratt answered it. and it does matter. bad decisions re this one issue results in failed and cancelled projects, late projects, budget overruns, etc.
curi:
lots of ppl's knowledge of decision making processes extends about as far as pro/con lists and ad hoc arguing.
curi:
so they are implicitly betting a significant amount of project effectiveness on something like "my foundation of pro/con lists and ad hoc arguing is adequate knowledge of decision making processes".
curi:
this is ... unwise.
curi:
another generic issue is lying. what is a lie? how do you know when you're lying to yourself? a lot of ppl make a bet roughly like "either my standard cultural knowledge + random variance about lying is good or lying won't come up in the project".
curi:
similar with bias instead of lying.
curi:
another common, generic way projects go wrong is ppl never state the project goal. they don't have clear criteria for project success or failure.
curi:
related, it's common to make basically no attempt to estimate the resources needed to complete the project successfully and estimating the resources available and comparing those two things.
curi:
goals and resource budgeting are things some ppl actually do. they aren't rare. but they're often omitted, especially for more informal and non-business projects.
curi:
including some very ambitious change-the-world type projects, where considering a plan and what resources it'll use is actually important. a lot of times ppl do stuff they think is moving in the direction of their goal without seriously considering what it will take to actually reach their goal.
curi:
e.g. "i will do X to help the environment" without caring to consider what breakpoints exist for helping the environment that make an important difference and how much action is required to reach one.
curi:
there are some projects like "buy taco bell for dinner" that use low resources compared to what you have available (for ppl with a good income who don't live paycheck to paycheck), so you don't even need to consciously think through resource use. but a lot of bigger ones one ought to estimate e.g. how much time it'll take for success and how much time one is actually allocating to the project.
curi:
often an exploratory project is appropriate first. try something a little and see how you like it. investigate and learn more before deciding on a bigger project or not. ppl often don't consciously separate this investigation from the big project or know which they are doing.
curi:
and so they'll do things like switch to a big project without consciously realizing they need to clear up more time on their schedule to make that work.
curi:
often they just don't think clearly about what their goals actually are and then use bias and hindsight to adjust their goals to whatever they actually got done.
curi:
there are lots of downsides to that in general, and it's especially bad with big ambitious change/improve the world goals.
curi:
one of the most egregious examples of the broad issues i'm talking about is political activism. so many people are working for the red or blue team while having done way too little to find out which team is right and why.
curi:
so they are betting their work on their political team being right. if their political team is wrong, their work is not just wasted but actually harmful. and lots of ppl are really lazy and careless about this bet. how many democrats have read one Mises book or could name a book or article that they think refuses a major Mises claim?
curi:
how many republicans have read any Marx or could explain and cite why the labor theory of value is wrong or how the economic calculation argument refutes socialism?
curi:
how many haters of socialism could state the relationship of socialism to price controls?
curi:
how many of them could even give basic economic arguments about why price controls are harmful in a simple theoretical market model and state the premises/preconditions for that to apply to a real situation?
curi:
i think not many even when you just look at people who work in the field professionally. let alone if you look at people who put time or money into political causes.
curi:
and how many of them base their dismissal of solipsism and idealism on basically "it seems counterintuitive to me" and reject various scientific discoveries about quantum mechanics for the same reason? (or would reject those discoveries if they knew what they were)
curi:
if solipsism or idealism were true it'd have consequences for what they should do, and people's rejections of those ideas (which i too reject) are generally quite thoughtless.
curi:
so it's again something ppl are betting projects on in an unreasonable way.
curi:
to some extent ppl are like "eh i don't have time to look into everything. the experts looked into it and said solipsism is wrong". most such ppl have not read a single article on the topic and could not name an expert on the topic.
curi:
so their bet is not really on experts being right – which if you take that bet thousands of time, you're going to be wrong sometimes, and it may be a disaster – but their bet is actually more about mainstream opinion being right. whatever the some ignorant reporters and magazine writers claimed the experts said.
curi:
they are getting a lot of their "expert" info fourth hand. it's filtered by mainstream media, talking heads on TV, popular magazines, a summary from a friend who listened to a podcast, and so on.
curi:
ppl will watch and accept info from a documentary made by ppl who consulted with a handful of ppl who some university gave expert credentials. and the film makers didn't look into what experts or books, if any, disagree with the ones they hired.
curi:
sometimes the info presented disagrees with a majority of experts, or some of the most famous experts.
curi:
sometimes the film makers have a bias or agenda. sometimes not.
curi:
there are lots of issues where lots of experts disagree. these are, to some rough approximation, the areas that should be considered controversial. these merit some extra attention.
curi:
b/c whatever you do, you're going to be taking actions which some experts – some ppl who have actually put a lot of work into studying the matter – think is a bad idea.
curi:
you should be careful before doing that. ppl often aren't.
curi:
politics is a good example of this. whatever side you take on any current political issue, there are experts who think you're making a big mistake.
curi:
but it comes up in lots of fields. e.g. psychiatry is much less of an even split but there are a meaningful number of experts who think anti-psychotic drugs are harmful not beneficial.
curi:
one of the broad criteria for areas you should look into some before betting your project on are controversial areas. another is big risk areas (it's worse if you're wrong, like AI risk or e.g. there's huge downside risk to deciding that curing aging is a bad cause).
curi:
these are imperfect criteria. some very unpopular causes are true. some things literally no one currently believes are true. and you can't deal with every risk that doesn't violate the laws of physics. you have to estimate plausibility some.
curi:
one of the important things to consider is how long does it take to do a good job? could you actually learn about all the controversial areas? how thoroughly is enough? how do you know when you can move on?
curi:
are there too many issues where 100+ smart ppl or experts think ur initial plan is wrong/bad/dangerous, or could you investigate every area like that?
curi:
relying on the opinions of other ppl like that should not be your whole strategy! that gives you basically no chance against something your culture gets systematically wrong. but it's a reasonable thing to try as a major strategy. it's non-obvious to come up with way better approaches.
curi:
you should also try to use your own mind and judgment some, and look into areas you think merit it.
curi:
another strategy is to consider things that people say to you personally. fans, friends, anonymous ppl willing to write comments on your blog... this has some merits like you get more customized advice and you can have back and forth discussion. it's different to be told "X is dangerous b/c Y" from a book vs. a person where you can ask some clarifying questions.
curi:
ppl sometimes claim this strategy is too time consuming and basically you have to ignore ~80% of all criticism you're aware of with according to your judgment with no clear policies or principles to prevent biased judgments. i don't agree and have written a lot about this matter.
curi:
i think this kind of thing can be managed with reasonable, rational policies instead of basically giving up.
curi:
some of my writing about it: https://elliottemple.com/essays/using-intellectual-processes-to-combat-bias
curi:
most ppl have very few persons who want to share criticism with them anyway, so this article and some others have talked more about ppl with a substantial fan base who actually want to say stuff to them.
curi:
i think ppl should write down what their strategy is and do some transparency so they can be held accountable for actually doing it in addition to the strategy itself being something available for ppl to criticize.
curi:
a lot of times ppl's strategy is roughly "do whatever they feel like" which is such a bias enabler. and they don't even write down anything better and claim to do it. they will vaguely, non-specifically say they are doing something better. but no actionable or transparent details.
curi:
if they write something down they will want it to actually be reasonable. a lot of times they don't even put their policies into words into their own head. when they try to use words, they will see some stuff is unreasonable on their own.
curi:
if you can get ppl to write anything down what happens next is a lot of times they don't do what they said they would. sometimes they are lying pretty intentionally and other times they're just bad at it. either way, if they recognize their written policies are important and good, and then do something else ... big problem, even in their own view.
curi:
so what they really need are policies which some clear steps and criteria where it's really easy to tell if they are being done or not. just just vague stuff about using good judgment or doing lots of investigation of alternative views that represent material risks to the project. actual specifics like a list of topic areas to survey the current state of expert knowledge in with a blog post summarizing the research for each area.
curi:
as in they will write a blog post that gives info about things like what they read and what they think of it, rather than them just saying they did research and their final conclusion.
curi:
and they should have written policies about ways critics can get their attention, and for in what circumstances they will end or not start a conversation to preserve time.
curi:
if you don't do these things and you have some major irrationalities, then you're at high risk of a largely unproductive life. which is IMO what happens to most ppl.
curi:
most ppl are way more interested in social status hierarchy climbing than taking seriously that they're probably wrong about some highly consequential issues.
curi:
and that for some major errors they are making, better ideas are actually available and accessible right now. it's not just an error where no one knows better or only one hermit knows better.
curi:
there are a lot of factors that make this kind of analysis much harder for ppl to accept. one is they are used to viewing many issues as inconclusive. they deal with controversies by judging one side seems somewhat more right (or sometimes: somewhat higher social status) instead of actually figuring out decisive, clear cut answers.
curi:
and they think that's just kinda how reason works. i think that's a big error and it's possible to actually reach conclusions. and ppl actually do reach conclusions. they decide one side is better and act on it. they are just doing that without having any reason they regard as adequate to reach that conclusion...
curi:
some of my writing about how to actually reach conclusions re issues http://curi.us/1595-rationally-resolving-conflicts-of-ideas
curi:
this (possibility of reaching actual conclusions instead of just saying one side seems 60% right) is a theme which is found, to a significant extent, in some of the other thinkers i most admire like Eliyahu Goldratt, Ayn Rand and David Deutsch.
curi:
Rand wrote this:
curi:
Now some of you might say, as many people do: “Aw, I never think in such abstract terms—I want to deal with concrete, particular, real-life problems—what do I need philosophy for?” My answer is: In order to be able to deal with concrete, particular, real-life problems—i.e., in order to be able to live on earth.
You might claim—as most people do—that you have never been influenced by philosophy. I will ask you to check that claim. Have you ever thought or said the following? “Don’t be so sure—nobody can be certain of anything.” You got that notion from David Hume (and many, many others), even though you might never have heard of him. Or: “This may be good in theory, but it doesn’t work in practice.” You got that from Plato. Or: “That was a rotten thing to do, but it’s only human, nobody is perfect in this world.” You got it from Augustine. Or: “It may be true for you, but it’s not true for me.” You got it from William James. Or: “I couldn’t help it! Nobody can help anything he does.” You got it from Hegel. Or: “I can’t prove it, but I feel that it’s true.” You got it from Kant. Or: “It’s logical, but logic has nothing to do with reality.” You got it from Kant. Or: “It’s evil, because it’s selfish.” You got it from Kant. Have you heard the modern activists say: “Act first, think afterward”? They got it from John Dewey.
Some people might answer: “Sure, I’ve said those things at different times, but I don’t have to believe that stuff all of the time. It may have been true yesterday, but it’s not true today.” They got it from Hegel. They might say: “Consistency is the hobgoblin of little minds.” They got it from a very little mind, Emerson. They might say: “But can’t one compromise and borrow different ideas from different philosophies according to the expediency of the moment?” They got it from Richard Nixon—who got it from William James.
curi:
which is about how ppl are picking up a bunch of ideas, some quite bad, from their culture, and they don't really know what's going on, and then those ideas effect their lives.
curi:
and so ppl ought to actually do some thinking and learning for themselves to try to address this.
curi:
broadly, a liberal arts education should have provided this to ppl. maybe they should have had it by the end of high school even. but our schools are failing badly at this.
curi:
so ppl need to fill in the huge gaps that school left in their education.
curi:
if they don't, to some extent what they are at the mercy of is the biases of their teachers. not even their own biases or the mistakes of their culture in general.
curi:
schools are shitty at teaching ppl abstract ideas like an overview of the major philosophers and shitty at teaching practical guidelines like "leave 1/3 of your time slots unscheduled" and "leave at least 1/3 of your income for optional, flexible stuff. don't take on major commitments for it"
curi:
(this is contextual. like with scheduling, if you're doing shift work and you aren't really expected to think, then ok the full shift can be for doing the work, minus some small breaks. it's advice more for ppl who actually make decisions or do knowledge work. still applies to your social calendar tho.)
curi:
(and actually most ppl doing shift work should be idle some of the time, as Goldratt taught us.)
curi:
re actionable steps, above i started with addressing the risky bets / risky project premises. with first brainstorming things on the list and organizing into categories. but that isn't where project planning starts.
curi:
it starts with more like
curi:
goal (1 sentence). how the goal will be accomplished (outline. around 1 paragraph worth of text. bullet points are fine)
curi:
resource usage for major, relevant resource categories (very rough ballpark estimates, e.g. 1 person or 10 or 100 ppl work on it. it takes 1 day, 10 days, 100 days. it costs $0, $1000, $1000000.)
curi:
you can go into more detail, those are just minimums. often fine to begin with.
curi:
for big, complicated projects you may need a longer outline to say the steps involved.
curi:
then once u have roughly a goal and a plan (and the resource estimates help give concrete meaning to the plan), then you can look at risks, ways it may fail.
curi:
the goal should be clearly stated so that someone could clearly evaluate potential outcomes as "yes that succeeded" or "no, that's a failure"
curi:
if this is complicated, you should have another section giving more detail on this.
curi:
and do that before addressing risks.
curi:
another key area is prerequisites. can do before or after risks. skills and knowledge you'll need for the project. e.g. "i need to know how wash a test tube". especially notable are things that aren't common knowledge and you don't already know or know how to do.
curi:
failure to succeed at all the prerequisites is one of the risks of a project. the prerequisites can give you some ideas about more risks in terms of intellectual bets being made.
curi:
some prerequisites are quite generic but merit more attention than they get. e.g. reading skill is something ppl take for granted that they have, but it's actually an area where most ppl could get value from improving. and it's pretty common ppl's reading skills are low enough that it causes practical problems if they try to engage with something. this is a common problem with intellectual writing but it comes up plenty with mundane things like cookbooks or text in video games that provides information about what to do or how an ability works. ppl screw such things up all the time b/c they find reading burdensome and skip reading some stuff. or they read it fast, don't understand it, and don't have the skill to realize they missed stuff.)
curi:
quite a few writers are not actually as good at typing as they really ought to be, and it makes their life significantly worse and less efficient.
curi:
and non-writers. cuz a lot of ppl type stuff pretty often.
curi:
and roughly what happens is they add up all these inefficiencies and problems, like being bad at typing and not knowing good methods for resolving family conflicts, and many others, and the result is they are overwhelmed and think it'd be very hard to find time to practice typing.
curi:
their inefficiencies take up so much time they have trouble finding time to learn and improve.
curi:
a lot of ppl's lives look a lot like that.