Programming

14521 readers
1 users here now

All things programming and coding related. Subcommunity of Technology.


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS
1
 
 

cross-posted from: https://feddit.nl/post/45839000

When publishing a package for use by programmers, automated changelog generation is very beneficial. In this blog post, I explore how to do it in a simple way that works everywhere.

2
 
 

cross-posted from: https://lemmy.zip/post/52108691

Hi, this is a post for you to rant on your sore points on technology

See I am trying to think of a good project idea one that people actually want solved, is there an app you wished existed, a site u wanted, put it down here and hey what do you know you may just see an ad in some while that now it exists

3
4
5
6
7
 
 

What are some general recommendations to maintain a set of patches on top of a library distributed as a source tarball? Until now I've been adding the versions to a git repository by deleting previous files and adding the new files in a "upstream" branch, then merging that branch into the main branch which contains my patches. This turned into more work than I expected because they started moving files around and renaming in new releases.

I should probably be rebasing instead of merging, but are there any other recommendations for this type of situation?

8
 
 

cross-posted from: https://leminal.space/post/24911246

I'll be self-hosting a service with user submissions soon, so I'm worried about the https://howto.geoblockthe.uk/ situation.

Based on this I've wondered, are there any community maintained geo block lists that might be useful? All database options I found are either 1. an on-demand online service which seems questionable for privacy reasons, or 2. IPv4 only, or 3. have weird terms of use with a gag clause regarding the entire company making it and other weird stuff.

I'm not a fan of geo blocking in general, but the situation is what it is.

PS: Please don't discuss the Online Safety Act itself too much in the comments, or whether somebody should be using a geo ip to handle this. While I might appreciate useful input on that, I'm hoping this post can remain a resource for those who are looking for such a database for other reasons as well.

9
10
 
 

TypeScript does not throw an error at compile time for accessing an out-of-bounds index. Instead, it assumes that the value could be one of the types defined in the array (in this case, 1 or 2) or undefined.

TypeScript automatically infers the type of a value accessed from an array, even if that access is out of bounds. It assumes that the value could be one of the defined types or undefined, which can lead to confusion if you expect stricter enforcement of valid indices.

I just spent the last 2 hours trying to understand why I was getting a valid type from something that shouldn't have been valid.

I think that the hate that JavaScript receives is well deserved, at least coming from Rust this is an absolute nightmare.

11
10
Why Javascript? (feddit.it)
submitted 4 months ago* (last edited 4 months ago) by dontblink@feddit.it to c/programming@beehaw.org
 
 

After a while since I left JS, I decided to learn Typescript, React and to give it a go again.. It's all fun, until you have to hunt down bugs..

I'm working with an headless CMS, so you have like 4 layers of stuff: JS, TS, React, CMS.

My opinion is that debugging works for SIMPLE stuff, like if I have a static website generator, debugging and hunting the errors by hand to find what's wrong it's okay. But when you have to do really complex stuff, interact with APIs, working with promises, async requests and all, I do think that hunting bugs like that is the recipe for a disaster.

My project runs into a bug: I get 27 different errors messages from different tools, some of them are easy to misunderstand and not really pointing out the right problem. And I don't want to imagine what's like working with a big fullstack app where you have to manage backend as well, not just its APIs..

Once things starts to get complex the language MUST have a good compiler, MUST have a very good package manager, an included linter which is the SAME one for everyone, and API/libraries documentation tool which looks everywhere the same, and most importantly a carefully crafted error handling system. Once a language is very complex it needs to goes straight to the point and tells you where the issue is, what you are doing wrong, and it's okay to have layers but you cannot use 5 different error detection mechanisms for one project.

12
 
 

I have heard about Elninki's and Harvard's programs, can I join them now or is there a specific date?

13
 
 

Consider I only know apis are structured data that can be called or modified from within a program, and have no real further knowledge in real use cases nor in networking.

Where should I start from? Should I study backend?

I prefer docs rather than videos.

14
 
 

Need a quick comment

Recently just developed a tool/utility for windows, are there many free publishing sites?, I know about softpedia and apparently despite the age old UI its quite active!

Anyways just please take a moment and comment active software sites where I could publish

Here is my softpedia page if you have any specific advice regarding my particular project: https://www.softpedia.com/get/Security/Password-Managers-Generators/Wifi-Password-Finder.shtml

15
16
17
 
 

I want to have a mirror of my local music collection on my server, and a script that periodically updates the server to, well, mirror my local collection.

But crucially, I want to convert all lossless files to lossy, preferably before uploading them.

That's the one reason why I can't just use git - or so I believe.

I also want locally deleted files to be deleted on the server.

Sometimes I even move files around (I believe in directory structure) and again, git deals with this perfectly. If it weren't for the lossless-to-lossy caveat.

It would be perfect if my script could recognize that just like git does, instead of deleting and reuploading the same file to a different location.

My head is spinning round and round and before I continue messing around with find and scp it's time to ask the community.

I am writing in bash but if some python module could help with it I'm sure I could find my way around it.

TIA


additional info:

  • Not all files in the local collection are lossless. A variety of formats.
  • The purpose of the remote is for listening/streaming with various applications
  • The lossy version is for both reducing upload and download (streaming) bandwidth. On mobile broadband FLAC tends to buffer a lot.
  • The home of the collection (and its origin) is my local machine.
  • The local machine cannot act as a server
18
19
 
 

Hello! Basically, I need to process a very large (4000 lines) file and free ai chatbots like chatgpt aren't able to handle it. I would like to split it into smaller parts and process each part separately. I'm having however a very hard time finding a chatbot with free API. the only one I found is huggingchat, but after a few requests waiting 1 seconds before sending the next one it starts giving rate limit errors.

any suggestion? thanks in advance!

EDIT: I also tried to run gpt4all on my laptop (with integrated graphics) and it took like 2-5 minutes to asnwer a simple "hello" prompt, so it's not really feasable :(

20
 
 

Hi! I've been working with Hugo for a while and I also created a free MIT licensed theme with it!. I love the flexibility and the ease of use.

But I'll have to wok on a bit more complicated project than a simple showcase website/blog. The content to be published on it is not a lot, but it would be definitely better if I could:

  • Get/Post some content with API to avoid posting multiple times the same articles on different platforms, getting modifications as well.
  • Send posts digests via email / Download PDF post digests.
  • Post on social medias (?)
  • Parse some content from CSV files / I don't know anything about databases.

Now I know that I can do something like this with a little systemd service I might write on my own and something like Zapier + RSS feed + Mailchimp. Also I could leverage Hugo modules and the .GetRemote / transform.unmarshal command, to get content from remote sources.

Now I'm not really a lot more than an amateur developer, I was thinking a headless CMS could pheraps do this stuff and more in a better way (?). I'm not a webdev and I know only really really basic JavaScript, I can use Bootstrap for frontend confidently and add SCSS to it. I know a bit of Rust too.

Would it be worth to take the time learning how headless CMS's work? I don't really want to go back managing Wordpress plugins, updates ecc.

Do you think I'm going out of a static site generator purpose with this kind of project?

21
22
23
24
 
 

Let's say I have to host 25 websites.. How do I know how powerful should my VPS be? Which specs it should have, how fast the connection should be to handle X visits per day?

How do you understand which are your system requirements BEFORE deploying a project? Do you just make estimates and then scale up? Or there's some kind of tool to benchmark? how to handle this kind of stuff?

25
view more: next ›