Skip to content

briandoyle81/adventOfCode2021

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

34 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

adventOfCode2021

This readme has spoilers, as of course, does the code within.

As with before, I might someday use this for students, so I'm not cleaning up my final answers. See inside for a window into how my brain works.

Unedited notes below:

01

...

06

This is the first one I wasn't able to do part 2 with my goal of < 5 minutes, < 5 new lines of code. I did brute force for part 1

07

First one wasn't too hard. Realized the median would be the middle of all distances total.

Part 2 is the first one I didn't get right on the first submission. I was thinking that since the median discounts outliers, and average includes it, that it would be a good approach to try for the second problem where costs increase. Didn't work, but it felt like it might be an off by 1 error or a rounding error.

I tested a -1, +1, and a few beyond in each direction, and -1 provided the lowest cost, so I submitted that and it worked. Not satisfied with this solution though.

08

Not very hard, but so very messy. Still, I like problems with very ugly solutions, although this one could use a lot of cleaning up.

I feel like there is a more elegant solution using a graph, similar to word ladder, but I couldn't find it.

09

Finally a graph problem! First part was straightforward. Correct first sub, much cleaner code than last time.

Second part more or less straightforward as well. Wrong first submission though...because I'm a dunce and didn't retest after I moved the last few lines into a function, and I dropped sorting the array of basin sizes in doing so.

10

Balanced brackets!

Had a gauntlet thrown down. Part 1 in 5 to 10 minutes or so!

Part 2 took a bit longer. Was rushing to compete with a friend and of course had lots of typos, etc. Initially thought that my solution was wrong, then I realized I didn't count the score from the right direction.

11

This one was ugly. Started after driving home after a long weekend. Spent forever fighting through minor errors in my approach for part 1.

But I got part 2 in less than 5 minutes and less than 5 lines of code!

12

This one was a ton of fun for part 1. Knew it would be a traversal. Took a little bit to hone it to produce the correct result.

Part 2 was straightforward but took 30 lines and 30 minutes, mostly because I have a hard time with complicated logic statements.

13

Pretty simple for part 1, part two was fast and a ton of fun! Very clever!

14

Part 1 is a simple string manipulation problem. I need to circle back and figure out an elegant way to find the least value while counting.

Stuck for the first time on part 2 of this one. After a good bit of thinking, I thought maybe it might be made better by memoizing already calculated sub-patterns. It appears to be, but it's still, very, very slow, and I don't think it will get to the end in doable time.

Computer is chugging right now on step 30.

Not possible to calculate the polymer. Asked for and received a hint from a friend confirming.

So how to calculate the result each step without calculating the polymer...

With a second hint, reminding me of what I know at the start, what the pairs are, the starting polymer, and the ruleset, I picked up on keeping track of the number of each pairs and built a solution that way.

TODO: I'm annoyed that the solution cheats and doesn't actually find what you need in the story. I should try to find that polymer...

15

Today I remembered that you can put a tuple in a set 🤦

Anyhow, got to do Djikstra for the first time in awhile. I like that one, got me a job once. But I haven't done it in years, so I looked up the algo and implemented it from the description.

Took longer than I would have liked to put the expanded data set in memory, particularly since I don't think I can use it.

In the meantime, letting Djikstra run. 250000 nodes in linear time will only take an hour or so.

Culling options above and left of each step make it much faster, but still slow.

I think I could improve this by tracking if I went down or right and only adding a row or column.

Nope, another wrong answer. It must be able to go back up. Or not, misread a test on the small data.

Second answer wrong as well. Something wrong with the last step of the heap?

Ah, I misread the heap docs about what replace does. I needed to implement a priority queue as described in the docs for heapq.

Tough, and wrong entries, but I didn't need help, just research to relearn Dijkstra and the docs for Pythons built-in heap.

16

So far this one reads as though the challenge is how to write code when the person who gave you the spec is gaslighting you. I think they accidentally a sentence or two.

Pretty easy once the massive reading comprehension effort was done.

Bad submission for part because I made the bad choice to not fix the bad switch implementation I found online.

17

This one is a ton of fun! But I'm not sure how you can exclude a ceiling for velocity. It seems to be that with infinitely increasing Y, even though the velocity from gravity increases makes a miss more likely, I don't see how with infinite height you can't always find a higher hit.

I tested out to Y velocity of 10000 for part 1, which was slow, and I wasn't confident when I submitted. Right the first time, but not confidently.

My try everything algo is missing two hits in the first try for part 2...

Debugging by comparing to the results data given.

{(7, -1), (6, 0)} are misses for some reason, testing manually.

Got lucky here! Was comparing Y misses against the higher Y range, not lower!

Got part 2 on the first submission as well, but I'm really not 100% satisfied that there aren't a few higher options up somewhere...

18

This one has been pretty terrible. Explode seems like it should be simple, but it's completely awful.

I'm going to have to abandon this one for now.

Couldn't stay away, revisited string explode attempt. Seemed to get it to work, it handles the first addition case fine, and gets a result, but the result is wrong for the first addition problem in the small and regular test cases.

Good news, my explode finally seems to be working. It's a mess, but it works.

Bad news, my beautiful, elegant, intuitive, recursive split doesn't work. It splits them all at once. Which makes me wonder if a previous iteration of explode worked?

In any case, that was easy to fix, but my elation has turned quickly to horror in remembering this is only part 1.

Mercifully, the stupid, brute force way worked! About 5 lines, < 5 minutes to implement.

Success!?!?!

19

A very fun but very hard 1. For part 1 I did optimized brute force. Pre-calculate what I could and store it in sets for fast lookup. The result is a pretty bad polynomial complexity, but doable with this data size.

For part two, the first part was to store the scanner positions, which is just the offsets in part 1. Then it's easy just to do an n^2 comparison and find the longest.

20

This was a ton of fun. I did the virtual Conway solution like last year, but got my first submission wrong because it was hard to handle the inversion in the real enhancement algorithm that isn't in the test data set.

I ended up solving it by inverting the "lit" pixels every other step to avoid adding infinity to the lit set.

Aced Part II!!! Just needed to change the parameter from 2 to 50 to get it!

21

Part 1 was super easy. Took maybe 20 minutes. This is scary!

Oh my part 2 :D

It actually didn't end up being too bad. I had the approach right away of counting in groups of similar games and calculating by the possible roll totals and how many combos for each, but my first pass was too slow because I treated each combo of position and score as unique (well, I forgot to track position at first then added it in). By nesting scores inside position, I got it to be very fast.

22

I think part 1 should be fairly easy by using a virtual grid approach like before. My guess is that part 2 will be oh, don't ignore after 50, it's 50k big.

Oh, actually that might not work. The last two steps in the example complicate things.

Went ahead and did part 1 with the set of coordinates. It was a little slow, probably won't work for part 2, even though the time complexity is linear for N of total number of cubes changed.

It did not work for part 2. Went with an OOP method tracking cubes, and then sub-cubes removed from those cubes.

It was much faster, but I spent hours chasing an bug that made me doubt reality. It was a copy/paste error. I should really listen to my own advice more...

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages