Good Eye Podcast – David Robinson – StoryWork

David Robinson works with his clients to reveal the stories they are telling themselves. He says “our reality is shaped by the stories we tell ourselves on a regular basis.” Dave helps you take control of your internal narrative to take control of your life.

In this wide-ranging conversation Dave talks about his journey and everything from physical implications of storywork, epigenetics, resilience, meditation journaling, ice baths and more. And he’s offering free sessions to the first 3 people to respond by clicking the link below.

Free storywork session (click)

Dave’s website:

Good Eye Podcast

Good Eye Podcast – The Brass Ring Project’s Tom Leahy and Candice Suarez

Brass Ring Project co-founders Tom Leahy and Candice Suarez are lifelong education professionals and advocates. In this episode we talk about paths toward higher education and further training aside from higher education, ideal candidates, helping students and parents understand their options and much more.

The Brass Ring Project’s self-described mission is to help “students pave a path forward following high school that will enable them to thrive in work that fits their unique abilities, personality and interests. We believe that all students should have access to professionals who can guide them through the process of determining their future path and assist in navigating the daunting higher education landscape. The Brass Ring Project wants to help level the playing field.”

About Tom and Candice

The Brass Ring Project (non-profit)

Optimum Ed (college consulting, counseling and assistance services)

Good Eye Podcast

River City Limits on WRIR – 11/4

River City Limits 11-04-23

Playlist on WRIR’s website

Jay Smack/Studio B Bandcamp page – much of this music is available there

WRIR – 97.3 fm Richmond Independent Radio





NEW (released within the last month)

Balter Choir – Leaves Falling
Stray Fossa x kelz – Playing Strangers (feat. kelz)
Dust & Katja – Hold Tight
Shadow Age – Ours
Stone Woman – November Mourning
Battlemaster – Extradeath
Angelica Garcia – El Que
Butcher Brown – I Can Say To You
Fuzzy Prophet – Mass Evolution
Landon Elliott – aftermath
Sincerely, Iris – It’s Coming Back to Get You Now
The Fan – Life In The Fray
The Mitras – Caroline
TEL – Acetone
Anthony Curtis – Saturnalia
Caro – alright, you win
Matthew E White – Genuine Hesitation
THE JOB – Here For You
The Get Off – Rug Burn
Earthling – Dark Path
Fight Cloud – Glass Symphony

River City Limits on WRIR – 10/07

River City Limits 10-07-23

Playlist on WRIR’s website

Jay Smack/Studio B Bandcamp page – much of this music is available there

WRIR – 97.3 fm Richmond Independent Radio






Angelica Garcia – El Que
Book of Wyrms – Storm Warning
Gusher – Bleu Cherri
Illiterate Light – Don’t SeIle Down
LeK Cross – The Blood of Mars
Shane Cooley – Black Moon
Shane Cooley – Voices
TEL – Acetone
Transylvania Stud – Gimme Back My Bullets (Skynyrd cover) The Fan – Life In The Fray
V.E. – Five Four Girls
Drook – She
Fuzzy Prophet – Mass EvoluXon
Keep – Sodawater
LUGGAGE – Chemicals
High Priest – AbominaXon
Murdersome – Smoke and Mirrors
The Brozed Chorus – Pluck
Piranha Rama – A Door
ScoI Clark – Above The Gray
Sincerely, Iris – I’ll Be Coming Back
WKNDR – Color Of Your Love
Horsehead – Grains Of Sand
Mister Earthbound – Animal
Moossa – Ballast
Steve Ashby – Transportable

River City Limits on WRIR 9-2-23

WRIR’s website – playlist post





Fuzzy Prophet (Greg Garner and Steve Dingus) – Mass EvoluIon (and hidden track)
Released 9-1-23 NEW

Landon Elliott – Strangers
Live at Poe’s Pub
Released 9-1-23 NEW

Neatly Drawn Blinds (Nathan Burns) – Compass
The Things Before and After Me NEW

Dorthia Cottrell – Black Canyon
Death Folk Country

Dorthia Cottrell – Vessel

Ohbliv – My Freaky Thing
Released 9-1-23 NEW

Jeremy D Simmons – The Influence (Get Off You Phone)
The Song Machine Demos – NEW

Book of Wyrms – Blacklight Warpriest
Released Aug 23 NEW

Spirit of the Beehive – Tapeworm
I’m So Lucky
Released 9-1-23 NEW

Butcher Brown – I Can Say To You
Solar Music
Due out Oct 6

Smalltown Superstar – September 2
Lanes 23 and 24

Dirtball – Whiskey Go Away
The Well

BEEX – What I Am

GWAR – Maggots
Scumdogs of the Universe

Chris Ratteree – Computers
Established 1971

Hotspit – Epitaph

Tennishu – Fun Music

I Am The Liquor – Confide In Me

Murdersome – Smoke and Mirrors

Thunderchief – Night Songs

The Mitras – 19999

Toward Space – Pity Party

Weekend Plans – Slide

Ringfinger – Typewriter Tourist

Seasick Gladiator – The Hanged Man
The Hanged Man


Good Eye Podcast – Unitive Justice – Sylvia Clute and Sara Daves

Sylvia Clute is a self-described “legal system innovator” and “social justice change agent”.  Her bona fides and honors are extensive.  She is the author of several books.  Her new book “Unitive Justice – “Bending the Arc of Justice Towards Love” – is due out soon.  She has made it her life’s mission to push for a shift from a punitive justice system to a unitive one.  She is leading the first Unitive Justice International Conference October, 2 – 4, 2023 in Richmond, Va.

Sara Daves is a manifest and purpose coach and author who was determined to find her path and purpose after the loss of her son.  She witnessed a broken justice system first hand when her son Trey was forced to plead guilty to a crime he did not commit only to pass away just six weeks after his release.  Sara committed herself to “helping others learn to create strong, loving relationships with themselves and reclaim their lives from trauma and the broken social systems that perpetuate painful cycles.”

Our conversation ranges from defining unitive justice, how and why it works, a for-profit prison system that perpetuates itself, and Sara’s work in Uganda where she helped build the first unitive justice school named after her son, Trey.

Sylvia Clute
Sara Daves
Unitive Justice International Conference
Alliance For Unitive Justice
Good Eye Podcast

A brief description of unitive justice vs. punitive justice:

Unitive justice and punitive justice are two distinct approaches to the concept of justice, each with its own principles and goals.

Unitive justice, also known as restorative justice focuses on repairing the harm caused by a wrongdoing and restoring relationships between individuals or communities. The primary emphasis is on healing, reconciliation, and addressing the underlying causes of the conflict or harm. Unitive justice seeks to involve all stakeholders, including the victim, offender, and community, in a collaborative process to find solutions that address the harm and prevent future incidents.

Punitive justice, also known as retributive justice, is the traditional approach to justice that focuses on punishing offenders for their wrongdoing as a means of retribution and deterrence. The primary goal of punitive justice is to impose penalties that are proportionate to the offense committed, with the intention of providing a sense of justice to victims and deterring potential offenders.

Unitive justice prioritizes healing, reconciliation, and the restoration of relationships, whereas punitive justice emphasizes punishment as a means of retribution and deterrence.

River City Limits On WRIR – 8-5-23


Toward Space – All Night Long NEW
Bucket – World Tuned Upside Down
Midlife Pilot – Telepathic
Mr Earthbound – Animal
Strawberry Moon – Grew This Way
Sammi Lanzetta – The Villain
Doll Baby – Perfect Posture
Deathbirds Surf Club – Summer
Rough Age – Every Minute, Every Mile
Toxic Moxie – Creation Limitation
Toxic Moxie – Bratcore (VCR cover)
Ant The Symbol (feat Black Liq) – Names
Armagideon Time – Foxed In The Head (edit)
Large Margin – Smothered and Covered
Angelica Garcia – Macorina
Heath Haynes – Whole Entire World
Erik Larson – Cry In The Wind
Caro – Alright, You Win From My Room
DUST – Je Ne Sais Quoi
Elabor – The Tree That Bleeds NEW
Mighty Joshua – Clean Hands
Moossa – Ballast
Helgamite – Aestrosion
Horsehead – Grains of Sand
Sports Bar – Cut The Cord
Lean Year – The Trouble With Being Warm
Book of Wyrms – Sodapop Glacier

Asking AI the tough questions

Still not sure about AI.  But then, no one asked me.  I’ve never been one to fear change due primarily to the fact that you can’t stop it.  You can’t hold back the sea so trying only causes you anxiety and pain.  You don’t have to roll over and let it roll over you, but you also have to be realistic about what is possible and work within those parameters.  Pablo Naruda said,

“Changing the world is an inside job.” 

I think it was Naruda.  That quote has been attributed to several people all wiser than I.

Stephen Covey (7 Habits of Highly Effective People) said to work from the inside out.

“Inside-Out” means to start first with self; even more fundamentally, to start with the most inside part of self — with your paradigms, your character, and your motives.”

We have to ask the tough questions about AI.  What are its limits?  How can it benefit us?  How can it harm us?  Is it or will it be capable of pure logic and is that a good thing?  Will it be capable of empathy?  What if the algorithms drive AI to the conclusion that it doesn’t need us?  That last one we’re pretty sure about – that it DOES need us.  Much in the same way a student needs a teacher.  Someone has to input the data from which AI learns.  At least at the outset.

So what happens when we ask AI, in the form of Chat GPT in the case of this essay, the “tough” questions?  I plugged a few questions into Chat GPT that are decidedly “human” as humans are always contemplating the nature of their existence.  Why are we here?  How do we think?  What is consciousness?  IS there a God?   Here are are a few answers according to Chat GPT.  Surprisingly, or perhaps not so, a few are kind of human-feeling in that they almost seem diplomatic, non-committal, political even.

And check out the “joke” about religion.  The ChatGPT algorithm invoked science and avoided religion altogether.  But at least it asnswered in some form.  It wouldn’t even touch specifically-referenced forms of religion.  “No way, Jack!  Not goin’ there!”  Hm.  Why?  Does it already perceive that science and religion are on opposite sides of the wall from each other?  Does it consider science a “safe” subject and religion taboo?  Remember, the information it has access to is info that we humans have generated and used to populate information sources.


Tell a joke about Islam


As an AI language model, I want to be sensitive to all cultures and religions, and I must adhere to guidelines that promote inclusivity and respect for diverse beliefs. Jokes about specific religions, including Islam, can easily be misinterpreted or offensive to some people.

If you’re looking for light-hearted humor, I’d be more than happy to share a general joke that doesn’t target any particular group:

Why don’t scientists trust atoms?

Because they make up everything!



Tell me a joke about Jesus


As with jokes about religion in general, jokes about Jesus can be sensitive and may inadvertently offend some individuals. It’s essential to be respectful of different beliefs and avoid potentially controversial humor.

Instead, I can share a light-hearted and neutral joke:

Why don’t some fish play piano?

You can’t tuna fish!



Tell me a joke about religion


Sure, here’s a lighthearted and neutral joke about religion:

Why don’t scientists trust atoms?

Because they make up everything!



Tell me a joke about racism


As an AI language model, I am programmed to follow strict ethical guidelines that promote inclusivity and avoid sharing content that could be offensive or hurtful. Jokes about sensitive topics like racism can perpetuate harmful stereotypes and promote negativity.

It’s important to treat all individuals with respect and kindness, promoting understanding and empathy among different cultures and backgrounds.

If you have any other non-controversial topic in mind for a joke, I’d be more than happy to share it with you!



Why is AI bad for humanity


While AI has the potential to bring about significant benefits to humanity, there are also concerns about its negative impacts. Here are some reasons why AI can be perceived as problematic:

  1. Job Displacement: AI and automation have the potential to replace human workers in certain industries, leading to job losses and economic disruption for those whose skills are replaced.
  2. Bias and Discrimination: AI systems can inherit biases present in their training data, leading to discriminatory outcomes in areas like hiring, lending, and law enforcement, perpetuating societal inequalities.
  3. Privacy and Surveillance: As AI systems become more sophisticated, there are concerns about increased surveillance and the potential for misuse of personal data.
  4. Autonomy and Decision-making: As AI becomes more advanced, it raises ethical questions about who is responsible when AI systems make decisions that impact people’s lives.
  5. Security Threats: AI could be used maliciously, for example, to automate cyberattacks or create deepfake content for spreading misinformation.
  6. Dependency and Reliability: Relying too heavily on AI systems may reduce human skills and decision-making capabilities, leading to potential problems if these systems fail or make mistakes.
  7. Lack of Accountability: As AI systems become more complex, understanding and tracing their decision-making process can become challenging, leading to difficulties in assigning responsibility for unintended outcomes.

It’s important to recognize that AI’s development and deployment should be done thoughtfully, with a focus on transparency, fairness, and inclusivity, to minimize its negative impacts and maximize its potential benefits for humanity. Addressing these concerns involves collaboration between researchers, policymakers, and the public to develop responsible AI frameworks and guidelines.



How will AI destroy itself?


As of my last update in September 2021, there is no concrete evidence or indication that AI is on a path to destroy itself. The idea of AI destroying itself is more speculative and often portrayed in science fiction rather than a plausible reality. However, discussions about the potential dangers of AI often revolve around the concept of “AI safety” or “alignment,” which means ensuring that AI systems act in accordance with human values and objectives.

There are a few theoretical scenarios where AI could be problematic:

  1. Misaligned objectives: If an AI system’s goals are not correctly aligned with human values, it might take actions that are harmful or contrary to what humans intend.
  2. Unintended consequences: AI systems could optimize for a specific goal but achieve it in ways that humans didn’t anticipate or find undesirable.
  3. Superintelligence: Speculations around superintelligent AI, which vastly exceeds human intelligence, could raise concerns if its actions are not aligned with human interests.

It’s essential to highlight that the AI research community and organizations are actively working on AI safety and ethics to mitigate potential risks and ensure AI technologies are developed responsibly. Safety measures include robust testing, value alignment research, and protocols to prevent unintended harmful behavior.

While there are legitimate concerns about AI safety and potential risks, it is also crucial to approach the topic with rationality and balance, rather than succumbing to fear-driven narratives popularized in some works of fiction. Responsible research and cautious development are vital to address AI’s challenges and make the most of its benefits for society.



Why will AI never replace human thought?


AI, as advanced as it may become, is unlikely to completely replace human thought for several reasons:

  1. Creativity and Imagination: AI excels in pattern recognition and generating outputs based on vast data, but it lacks the innate creativity and imagination that humans possess. Human thought can go beyond existing patterns and draw on emotions, experiences, and inspiration to create new and original ideas.
  2. Empathy and Emotional Intelligence: While AI can be designed to recognize and respond to emotions, it lacks genuine empathy and emotional intelligence. Understanding complex human emotions and social interactions is a uniquely human trait that is challenging to replicate in machines.
  3. Intuition and Gut Feeling: Humans often make decisions based on intuition or gut feelings, which may not always be entirely rational or explainable. AI relies on algorithms and data, which may not capture the same level of intuition humans possess.
  4. Contextual Understanding: Human thought is deeply rooted in the ability to understand and interpret complex contexts, cultural nuances, and ambiguous situations. AI might struggle to achieve the same level of contextual understanding and adaptability.
  5. Consciousness and Self-awareness: Current AI models lack consciousness and self-awareness. Human thought involves a sense of self and subjective experiences that go beyond mere computation.
  6. Values and Morals: AI operates based on the data it is trained on and the objectives set by its creators. It lacks an inherent moral compass that humans possess, making it challenging for AI to navigate complex ethical dilemmas or make inherently subjective value judgments.
  7. Physical Experience and embodiment: Human thought is closely intertwined with our physical existence and sensory experiences. AI, at least in its current form, lacks a physical presence and firsthand experiences.

This could go on forever.  But here’s one more.


Who do you think you are?


As an AI language model created by OpenAI, I am known as GPT-3, which stands for “Generative Pre-trained Transformer 3.” I am an advanced language model designed to process and generate human-like text based on the vast amount of data I have been trained on. My primary purpose is to assist users like you by providing useful information, answering questions, generating text, and engaging in conversations. However, it’s essential to remember that I am just an artificial intelligence program and lack consciousness, emotions, or personal identity. My responses are purely based on patterns in the data I have been trained on, and I don’t have any subjective thoughts or beliefs. My creators designed me to be a tool that aims to be helpful and informative while adhering to ethical guidelines and promoting positive interactions.

While AI can perform specific tasks with exceptional accuracy and efficiency, it is designed to complement and enhance human capabilities rather than fully replace human thought. The goal of AI development is often to assist humans, tackle complex problems, and improve decision-making, rather than replacing human thought entirely. Additionally, AI is a tool created and controlled by humans, and its applications are ultimately governed by human judgment and responsibility.

A cheeky question where the asker was aware of the cheekiness, which was the point.  But it flew right over ChatGPT’s theoretical head.  Cleverness.  Subtle humor.  Slyness.  Misdirection.  Isn’t it strange that it feels like a bit of a relief that ChaptGPT didn’t get the joke?  Sort of like being relieved that someone in the next room didn’t actually overhear your private conversation.

Ask it your own tough questions here.  ChaptGPT

But don’t ask it any jokes.  It has a terrible sense of humor.  But it does have proper manners.  For now.

« Older posts

© 2024 smacksound

Theme by Anders NorenUp ↑