Autonomía digital y tecnológica

Código e ideas para una internet distribuida

Linkoteca. Archivo de navegación. Page 2


Captura de pantalla de survey.stackoverflow.co/2024

In May 2024, over 65,000 developers responded to our annual survey about coding, the technologies and tools they use and want to learn, AI, and developer experience at work. Check out the results and see what’s new for Stack Overflow users.

Captura de pantalla de platform.openai.com/tokenizer

OpenAI’s large language models (sometimes referred to as GPT’s) process text using tokens, which are common sequences of characters found in a set of text. The models learn to understand the statistical relationships between these tokens, and excel at producing the next token in a sequence of tokens.

You can think of tokens as the “letters” that make up the “words” and “sentences” that AI systems use to communicate.

A helpful rule of thumb is that one token generally corresponds to ~4 characters of text for common English text. This translates to roughly ¾ of a word (so 100 tokens ~= 75 words).

The process of breaking text down into tokens is called tokenization. This allows the AI to analyze and “digest” human language into a form it can understand. Tokens become the data used to train, improve, and run the AI systems.

Why Do Tokens Matter? There are two main reasons tokens are important to understand:

  1. Token Limits: All LLMs have a maximum number of tokens they can handle per input or response. This limit ranges from a few thousand for smaller models up to tens of thousands for large commercial ones. Exceeding the token limit can lead to errors, confusion, and poor quality responses from the AI.
  2. Cost: Companies like OpenAI, Anthropic, Alphabet, and Microsoft charge based on token usage when people access their AI services. Typically pricing is per 1000 tokens. So the more tokens fed into the system, the higher the cost to generate responses. Token limits help control expenses.

Strategies for Managing Tokens

Because tokens are central to how LLMs work, it’s important to learn strategies to make the most of them:

  • Keep prompts concise and focused on a single topic or question. Don’t overload the AI with tangents.
  • Break long conversations into shorter exchanges before hitting token limits.
  • Avoid huge blocks of text. Summarize previous parts of a chat before moving on.
  • Use a tokenizer tool to count tokens and estimate costs.
  • Experiment with different wording to express ideas in fewer tokens.
  • For complex requests, try a step-by-step approach vs. cramming everything into one prompt.

To somewhat expand on the earlier answers here, there are a number of details which are commonly overlooked.

  • Prefer subprocess.run() over subprocess.check_call() and friends over subprocess.call() over subprocess.Popen() over os.system() over os.popen()
  • Understand and probably use text=True, aka universal_newlines=True.
  • Understand the meaning of shell=True or shell=False and how it changes quoting and the availability of shell conveniences.
  • Understand differences between sh and Bash
  • Understand how a subprocess is separate from its parent, and generally cannot change the parent.
  • Avoid running the Python interpreter as a subprocess of Python.
Data Science at the Command Line

This thoroughly revised guide demonstrates how the flexibility of the command line can help you become a more efficient and productive data scientist. You’ll learn how to combine small yet powerful command-line tools to quickly obtain, scrub, explore, and model your data. To get you started, author Jeroen Janssens provides a Docker image packed with over 100 Unix power tools.

DK: ¿Tú qué piensas de la enseñanza religiosa en los colegios?JMP:

Uf. Hombre, yo es que soy muy tradicional, no en el sentido banal de la palabra, sino que yo me considero parte de una cadena, entonces yo creo que los seres humanos tienen que sentirse partícipes de una comunidad. Y lo que le da sentido a las comunidades en términos históricos es la religión, entonces yo creo que un español debe ser instruido en la religión católica.

DK: Un español musulmán, por ejemplo.JMP:

Un español…

DK: Un español puede ser musulmán, y protestante, y judío… musulmanes y judíos ha habido en España casi más años que católicos, ¿no?JMP:

Sí, pero todo nuestro arte, toda nuestra literatura… todo es católico.

DK: Entonces tú piensas que un español musulmán debería recibir clases de religión católica.JMP:

Sí, pero no en sentido catequético.

DK: Bueno, como se hace ahora en los colegios.JMP:

No, no, pero eso es una lección, un derecho de los padres a elegir la religión en la que desean que sus hijos sean formados.

DK: Ese derecho no existe en la religión musulmana, por ejemplo, mientras que para la católica sí es un derecho efectivo en todos los colegios.JMP:

Sí, pero bueno, también es una religión mayoritaria. Pero bueno, yo la pondría para todo el mundo.

DK: Obligatoria para todo el mundo.JMP:

Sí. Sí. Pero no en términos confesionales, en términos de instrucción, o sea, que la gente sepa qué es la eucaristía, que sepa lo que es la virginidad de María…

DK: Historia de las religiones…JMP:

Pero bueno, es que no es lo mismo.

DK: ¿De verdad crees en serio que está bien que le expliquen al hijo de un musulmán español lo que es la virginidad de la Virgen, la eucaristía y todo eso?JMP:

Sí, porque ese niño podrá leerse un auto sacramental de Calderón de la Barca y entenderlo. O podrá leerse un soneto de Lope de Vega y entenderlo. Es una forma de hacerlo partícipe de una tradición cultural, de una tradición artística, de una tradición… Yo creo que sí. Pero en esto no veamos un deseo de imponer religión o una determinada religión a la gente, no, es deseo de hacerla partícipe de nuestra historia.

A hedge fund is an investment fund where capital from various institutional and individual investors is combined and put toward a range of assets, including derivatives, stocks, bonds, commodities and foreign currencies, with the aim of optimizing returns.

Hedge funds, in general, operate as limited partnerships, professionally managed by fund managers who pool money from investors. However, participation in hedge funds, including those dealing with cryptocurrencies, is typically limited to high-net-worth individuals who can bear higher management fees and associated risks.

Crypto hedge funds generate revenue through an annual management fee, typically ranging from 1% to 4% of the invested amount. In addition, investors may also be obligated to pay a percentage of earned profits as performance fees to the managing team.

Crypto hedge funds employ a combination of systematic and discretionary investment strategies to effectively navigate the crypto landscape. The systematic approach relies on computer transaction processing models, offering a structured framework, reducing emotional influences and providing consistency. However, the risk lies in the potential vulnerability of these algorithms to unforeseen market conditions.

L’équipe des Imaginations Fertiles vous partage sa boîte à outils !
Retrouvez ici des documents vierges & disponibles pour vous aider dans le développement de votre projet.

Vous montez un tiers-lieu ? Vous questionnez vos usagers ? Vous animez des temps d’intelligence collective ? Vous souhaitez découvrir le design de service ?

L’équipe des Imaginations Fertiles peut vous accompagner dans vos projets, n’hésitez pas à nous faire part de vos besoins !

Outils et ressources pour les tiers-lieux qui développent une offre de formation professionnelle

PiNG est une structure aux mutliples facettes. À la fois association d’éducation populaire ressource sur les cultures numériques à l’échelle régionale et nationale; animatrice de deux tiers-lieux à Nantes (l’atelier partagé du Breil et le fablab Plateforme C), et organisme de formation, depuis plusieurs années.
C’est au croisement de ces facettes que cet article ressource va se concentrer, dans la perspective de fournir aux tiers-lieux en Pays de la Loire et à l’échelle nationale, un retour d’expérience orienté «opérationnel» sur des outils et des astuces permettant de développer une offre de formation ayant vocation à s’inscrire en conformité des exigences du référentiel national qualité en matière de formation professionnelle. En bref, si le mot «qualiopi» vous est familier, vous êtes au bon endroit pour glaner quelques ressources contexualisées avec une activité de tiers-lieu!

Escena de Gety Back

The Beatles: Get Back is a documentary television series directed and produced by Peter Jackson. It covers the making of the Beatles’ 1970 album Let It Be (which had the working title of Get Back) and draws largely from unused footage and audio material originally captured for and recycled original footage from the 1970 documentary of the album by Michael Lindsay-Hogg. The docuseries has a total runtime of nearly eight hours, consisting of three episodes, each of duration between two and three hours covering about one week each, together covering 21 days of studio time.

An oxygen bar is an establishment, or part of one, that sells oxygen for recreational use.

Oxygen bar guests pay about one U.S. dollar per minute to inhale a percentage of oxygen greater than the normal atmospheric content of 20.9% oxygen.

It has been claimed by alternative medicine that the human body is oxygen-deprived, and that oxygen will remove «toxins» and even cure cancer.

The medical profession warns that individuals with respiratory diseases such as asthma and emphysema should not inhale too much oxygen.[5] Higher than normal oxygen partial pressure can also indirectly cause carbon dioxide narcosis in patients with chronic obstructive pulmonary disease (COPD).

Escena de Solo, a Star Wars Story

George Lucas began developing a Han Solo prequel in 2012 and commissioned Lawrence Kasdan to write the screenplay. After Lucas sold Lucasfilm to The Walt Disney Company in 2012, Kasdan was hired to write Star Wars: The Force Awakens (2015), leaving his son Jonathan to complete the Solo script. Principal photography began in January 2017 at Pinewood Studios, with directors Phil Lord and Christopher Miller. Both were fired in June 2017 following creative differences with Lucasfilm, and Howard was hired as their replacement. Solo is one of the most expensive films ever made, with a budget of at least $275 million.

I love the WordPress Rest API and switching more and more from theme development to a headless WP approach, with an nice front-end framework. Right now I’m favoring Nuxt.js, which is build on Vue.js (check out wuxt, my very own dockerized nuxt/wp development environment).

For using WPs full strength with the Rest API I’ve collected/build a useful snippet library with WordPress Rest API extensions. I’ll try to maintain the following list as development goes on. All of the following extensions can be embedded in the functions.php file. If you wondering about the wuxt_ prefix, I’ve got the code from my Wuxt project and the prefix is as good as anyone.

Detecting the Safari browser: The user-agent of the Safari browser is “Safari”. This value is passed to indexOf() method to detect this value in the user-agent string. One additional check is required in the case of the Safari browser as the user-agent of the Chrome browser also includes the Safari browser’s user-agent. If both the user-agents of Chrome and Safari are in the user-agent, it means that the browser is Chrome, and hence the Safari browser value is discarded.

interact.js takes a slightly different approach compared to most drag and drop libraries. To give you as much control as possible, it tries to provide a simple, flexible API that gives you all the pointer event data you’ll need to move elements around.

The library doesn’t even do any moving at all! This is great because you decide exactly what feedback your users get. You can drag elements, draw on a canvas or (if you really need to) scroll the page.

Technically they’re all machine learning models, but checkpoints are usually referred to just as models. All models are static, meaning they only know what they were trained on. In order for them to learn something new, they have to be (re)trained again, also known as finetuning.

Their differences in very rough terms:

Checkpoints are the big models that make images on their own.

Loras and all their variations like Lycoris are «mini models» that plug into a checkpoint and alter their outputs. They let checkpoints make styles, characters and concepts that the base checkpoint they’re used on doesn’t know or didn’t know very well.

Hypernetworks are an older and not as good implementation of the same paper and same concept as Loras.

Textual Inversions are sort of bookmarks or compilations of what a model already knows, they don’t necessarily teach something new but rearrange stuff that a model already knows in a way that it didn’t know how to arrange by itself.

SD 1.x, SD 2.x and SDXL are both different base checkpoints and also different model architectures. Think of them maybe as non backwards compatible consoles if that’s easier to understand. SD 1.5 is say both the NES itself but also one game for it. All SD 1.x based models are compatible with SD 1.x Loras and models for extensions like ControlNet. SD 2.X is the SNES, it’s a different architecture so 1.x models won’t be compatible with it, same for SDXL if you say that’s like the N64.

ControlNet models are also machine learning models that inject into the Stable Diffusion process and control the denoising process, they’re used with an image made by preprocessor. That image is be used to guide and control that denoising process, also hence the name.

All these models come in safetensors format. Safetensors is the standard for machine learning models, since they only contain the necessary data for diffusion, hence their name. .ckpt is the old model format that’s outdated, because it’s an unsafe format that can contain and execute malicious code.

Pruned models are models that have unnecessary weights and data removed. Weights are part of what the model learned in order to denoise noise to make an image. Say a model has a weight for alien = 0.00000000000000000001, it’s so minimal that it won’t do anything, but it’s still taking up space. Now multiply that for a lot of more useless weights, pruning removes all of them so only the relevant data is left.

FP16 models are smaller than FP32 models because they have lower precision, it’s basically like calculating pi with less numbers, you’ll get close enough results, almost exactly the same in most cases, but still not as precise. Yet the images from FP16 models are also not that different or worse, and A1111 converts models to FP16 by default when loading them anyway, for faster speed.

To use a model, place them in their specific folder inside the UI you’re using. For A1111 they go in stable-diffusion-webui\models in self explanatory folders for Lora, etc. and in stable-diffusion-webui\models\Stable-diffusion for checkpoints.

Both checkpoints and Loras can be used either for poses or for styles, depending on what they were trained on. Pose Loras are a thing for example.

The reason for your results is because you’re using SD 1.5 as a checkpoint. The base SD 1.5 checkpoint is almost a year old and the Lora you’re using was trained using not only newer checkpoints, but also checkpoints trained better on anime. I recommend a newer anime checkpoint for the image you’re trying to make. The base model field on Civitai models is really more like base architecture. Also it’s recommended to not go below 512 in resolution for SD 1.x models and not below 1024 for SDXL models.

Each of the events were 4-nights. One was more “professional”, with people paying €1500-2500 for a ticket. The others were more lowkey, with tickets between €200-500. All of our events are priced on a sliding scale so people with more income contribute more, making it more accessible to people with low income. Co-living guests paid €200-350/week (all inclusive)..

Nati & I were left with a €5k profit for the co-living period, and about a €10k profit for the 4 events we facilitated. €15k for 3 months is not an amazing salary for two people, but considering all our living expenses were covered, and the work is satisfying, it’s quite doable.

Next to the dinner table was the “community mastery board” – a facilitation tool we learned from our friend Drew from Agile Learning Centers. This is an easy way for people to name small problems and quickly come up with solutions together, like “the downstairs bedrooms have poor sound insulation” → “let’s keep the noise level down in that part of the house between 10pm and 8am”

If you’re an aspiring community-builder: I would definitely recommend you run an experiment something like this. But I should note there are other factors that enabled us to be successful.

For one: reputation. We’re well connected in multiple networks: TPOT, The Hum, Microsolidarity, and Enspiral. From my position in that web, it’s pretty easy to get an invitation in front of a few thousand people, so that takes most of the effort out of event promotion. 96% occupancy across 4 events is exceptionally high, and it’s the result of a decade of network-building internationally.

Another factor: skills for living in community.

Stable Diffusion images per minute. Benchmarking 50 GPUs

We’ve benchmarked Stable Diffusion, a popular AI image generator, on the 45 of the latest Nvidia, AMD, and Intel GPUs to see how they stack up. We’ve been poking at Stable Diffusion for over a year now, and while earlier iterations were more difficult to get running — never mind running well — things have improved substantially. Not all AI projects have received the same level of effort as Stable Diffusion, but this should at least provide a fairly insightful look at what the various GPU architectures can manage with AI workloads given proper tuning and effort.

The easiest way to get Stable Diffusion running is via the Automatic1111 webui project. Except, that’s not the full story. Getting things to run on Nvidia GPUs is as simple as downloading, extracting, and running the contents of a single Zip file. But there are still additional steps required to extract improved performance, using the latest TensorRT extensions. Instructions are at that link, and we’ve previous tested Stable Diffusion TensorRT performance against the base model without tuning if you want to see how things have improved over time. Now we’re adding results from all the RTX GPUs, from the RTX 2060 all the way up to the RTX 4090, using the TensorRT optimizations.

For AMD and Intel GPUs, there are forks of the A1111 webui available that focus on DirectML and OpenVINO, respectively. We used these webui OpenVINO instructions to get Arc GPUs running, and these webui DirectML instructions for AMD GPUs. Our understanding, incidentally, is that all three companies have worked with the community in order to tune and improve performance and features.

Portada del Calendario zaragozano de 2017

El calendario Zaragozano es una publicación anual española que incluye una predicción meteorológica-astronómica no científica del tiempo para un año. El contenido del pequeño calendario viene indicado por el siguiente subtítulo: “Juicio universal meteorológico, calendario con los pronósticos del tiempo, santoral completo y ferias y mercados de España”.

Physics has empiricism. If your physical theory doesn’t make a testable prediction, physicists will make fun of you. Those that do make a prediction are tested and adopted or refuted based on the evidence. Physics is trying to describe things that exist in the physical universe, so physicists have the luxury of just looking at stuff and seeing how it behaves.

Mathematics has rigor. If your mathematical claim can’t be broken down into the language of first order logic or a similar system with clearly defined axioms, mathematicians will make fun of you. Those that can be broken down into their fundamentals are then verified step by step, with no opportunity for sloppy thinking to creep in. Mathematics deals with ontologically simple entities, so it has no need to rely on human intuition or fuzzy high-level concepts in language.

Philosophy has neither of these advantages. That doesn’t mean it’s unimportant; on the contrary, philosophy is what created science in the first place! But without any way of grounding itself in reality, it’s easy for an unscrupulous philosopher to go off the rails. As a result, much of philosophy ends up being people finding justifications for what they already wanted to believe anyway, rather than any serious attempt to derive new knowledge from first principles. (Notice how broad the spread of disagreement is among philosophers on basically every aspect of their field, compared to mathematicians and physicists.)

This is not a big deal when philosophy is a purely academic exercise, but it becomes a problem when people are turning to philosophers for practical advice. In the field of artificial intelligence, things are moving quickly, and people want guidance about what’s to come. Should we consider AI to be a moral patient? Does moral realism imply that advanced AI will automatically prioritize humans’ best interests, or does the is-ought problem prevent that? What do concepts like «intelligence» and «values» actually mean?

Sobre la Gestión del conocimiento en las organizaciones:
Las tres leyes de la dinámica del conocimiento (Debra Amidon)
1. Cuando se comparte conocimiento, este se multiplica.
2. El valor se crea cuando el conocimiento se mueve desde un punto de origen a otro punto en el que existe una oportunidad o necesidad.
3. El apoyo mutuo con conocimiento proporciona la utilización óptima de los recursos, tanto tangibles como intangibles.

Herramientas citadas:
https://alternativeto.net/
https://www.zotero.org/
https://www.lesbonscomptes.com/recoll/pages/index-recoll.html
https://calibre-ebook.com/
https://gephi.org/users/download/
https://en.wikipedia.org/wiki/Zettelkasten
https://maadix.net/en/
https://latempesta.cc/es.html

Carlo Padial y Kike G de la Riva explicando mejor que nadie a donde viene y a donde va Internet
https://www.youtube.com/watch?v=c7dylxtVma4

Portada de La isla de las mujeres

Un piloto de acrobacias deseado por todas las mujeres, un pervertido incombustible, un aviador francés en plena I Guerra Mundial y un pobre diablo abatido y naufragado en una isla «aparentemente» desierta…A partir de aquí, este cómic dará un vuelco a todo lo creíais que era…

Página de La isla de las mujeres

macos-guest-virtualbox.sh is a Bash script that creates a macOS virtual machine guest on VirtualBox with unmodified macOS installation files downloaded directly from Apple servers.

A default install only requires the user to sit patiently and, less than ten times, press enter when prompted by the script, without interacting with the virtual machine.

Tested on bash and zsh on Cygwin. Works on macOS, CentOS 7, and Windows on x86 CPUs with VT-x or AMD-V. Should work on most modern Linux distros.

macOS Catalina (10.15), Mojave (10.14), and High Sierra (10.13) currently supported.

Jango Fett en El ataque de los clones

Watch the seeds of Anakin Skywalker’s transformation take root in Star Wars: Episode II – Attack of the Clones. Ten years after the invasion of Naboo, the galaxy is on the brink of civil war. Under the leadership of a renegade Jedi named Count Dooku, thousands of solar systems threaten to break away from the Galactic Republic. When an assassination attempt is made on Senator Padmé Amidala, the former Queen of Naboo, twenty-year-old Jedi apprentice Anakin Skywalker is assigned to protect her. In the course of his mission, Anakin discovers his love for Padmé as well as his own darker side. Soon, Anakin, Padmé, and Obi-Wan Kenobi are drawn into the heart of the Separatist movement and the beginning of the Clone Wars.

Cartel Star Wars 2, Attack of the Clones

Toni Segarra está considerado como el mejor creativo publicitario español del XX. Creador de anuncios cómo el de “Te gusta conducir” de BMW, “Bienvenido a la república independiente de tu casa” de Ikea o el de “¿A qué huelen las nubes?” de Evax. Durante su carrera ha ganado 39 leones en el Festival de Cannes y más de 100 soles en el Festival de San Sebastián. Ha trabajado en muchas agencias de publicidad como Vizeversa o Contrapunto y después montó la suya SCPF con varios amigos. La dejo en 2017 y decidió fundar Alegre Roca para ganar libertad y poder hacerse un poco más freelance.

Couverture de Saint-Elme 4

Gregor Mazur et son entourage débarquent à Saint-Elme pour l’enterrement de Roland Sax. Quatrième tome spectral de la saga de Lehman et Peeters.

La mort de Roland Sax crée des remous : Gregor Mazur et son entourage débarquent à Saint-Elme pour assister à l’enterrement. À la Vache Brûlée, Franck se remet de ses blessures et tente de convaincre Philippe de rester en ville malgré le danger. Stan et Yérim font un tour en boîte de nuit. Yves Mertens cherche une maison à louer. Et Piotr est touché par la grâce.