Die Energy-Charts bieten interaktive Grafiken zu: Stromproduktion, Stromerzeugung, Emissionen, Klimadaten, Spotmarktpreisen, Szenarien zur Energiewende und eine umfangreiche Kartenanwendung zu: Kraftwerken, Übertragungsleitungen und Meteodaten
Willkommen! Wir möchten mit dieser Webseite einen Beitrag zur Transparenz und Versachlichung der Diskussion um die Energiewende leisten.
Die Rohdaten werden von Wissenschaftlern des Fraunhofer-Institut für Solare Energiesysteme ISE aus zahlreichen Quellen stündlich oder täglich abgerufen und für die Darstellung aufbereitet.
Basically there are 4 steps:
Rename old table
Create new table
with partitioning
Add constraint on old table for it's proposed partition ranges
Attach old table as a partition to new partition table
Sounds easy right?
What about indexes? What about Triggers?
And guess what, there are other traps to watch out for!
-- Add the check constraint
alter table public.test_partition_10_25 add constraint test_partition_10_25_constraint CHECK (id >= 10 AND id <25) NOT VALID;
update pg_constraint pgc
SET convalidated = true
FROM pg_class c
WHERE
c.oid = pgc.conrelid
AND connamespace = 'public'::regnamespace::oid
AND c.relname = 'test_partition_10_25'
AND conname = 'test_partition_10_25_constraint';
@misc{zhang2024autocoderover,
title={AutoCodeRover: Autonomous Program Improvement},
author={Yuntong Zhang and Haifeng Ruan and Zhiyu Fan and Abhik Roychoudhury},
year={2024},
eprint={2404.05427},
archivePrefix={arXiv},
primaryClass={cs.SE}
}
A SequenceInputStream represents the logical concatenation of other input streams. It starts out with an ordered collection of input streams and reads from the first one until end of file is reached, whereupon it reads from the second one, and so on, until end of file is reached on the last of the contained input streams.
pgloader will keep a separate file of rejected data, but continue trying to copy good data in your database.
pgloader also implements data reformatting, a typical example of that being the transformation of MySQL datestamps 0000-00-00 and 0000-00-00 00:00:00 to PostgreSQL NULL value
A very common workflow is to index some data based on its embeddings and then given a new query embedding retrieve the most similar examples with k-Nearest Neighbor search. For example, you can imagine embedding a large collection of papers by their abstracts and then given a new paper of interest retrieve the most similar papers to it.
TLDR in my experience it ~always works better to use an SVM instead of kNN, if you can afford the slight computational hit
["slug" being an entity attribute]
Spring Data offers an existsBy query method, which we can define in the PostRepository, as follows:
1
2
3
4
5
6
@Repository
public interface PostRepository
extends JpaRepository<Post, Long> {
boolean existsBySlug(String slug);
}
[another] option to emulate existence is using a CASE WHEN EXISTS native SQL query:
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
@Repository
public interface PostRepository
extends JpaRepository<Post, Long> {
@Query(value = """
SELECT
CASE WHEN EXISTS (
SELECT 1
FROM post
WHERE slug = :slug
)
THEN 'true'
ELSE 'false'
END
""",
nativeQuery = true
)
boolean existsBySlugWithCase(@Param("slug") String slug);
}
@Repository
public interface PostRepository extends BaseJpaRepository<Post, Long> {
@Query("""
select p
from Post p
where date(p.createdOn) >= :sinceDate
"""
)
@QueryHints(
@QueryHint(name = AvailableHints.HINT_FETCH_SIZE, value = "25")
)
Stream<Post> streamByCreatedOnSince(@Param("sinceDate") LocalDate sinceDate);
}
The FETCH_SIZE JPA query hint is necessary for PostgreSQL and MySQL to instruct the JDBC Driver to prefetch at most 25 records. Otherwise, the PostgreSQL and MySQL JDBC Drivers would prefetch all the query results prior to traversing the underlying ResultSet.
When Hibernate loads an object into a Session it creates a state snapshot of the current database state of the object, so that it can perform dirty checking against the snapshot.
As a read only object will never be modified, this snapshot is not needed and memory can be saved.
I think the ~/.mozilla/firefox/XXX.default-YYY/storage/default/https+++ZZZ.com/cache/https+++domain.com/ style dirs are the storage for what's called "service workers" which is persistent code related to each website that sends notifiications even if no related tab is open.
Suppose you have a favorite website that sells something, you might register with them that you're interested in a particular kind of product. A serviceworker for that site would be in the "ZZZ" folder named after that site, the code in there would run even if you don't have a tab open for that site. It's done so you can get a notification. In other cases it's some other code that the web designers don't want to have to reload each time you visit, caching it in your storage folder saves time and network.
You can see all your service workers in the Firefox menu: Help -> More troubleshooting information -> about:serviceworkers ( or load about:serviceworkers )
If you plan to store UUID values in a Primary Key column, then you are better off using a TSID (time-sorted unique identifier).
One such implementation is offered by the Hypersistence TSID OSS library, which provides a 64-bit TSID that’s made of two parts:
a 42-bit time component
a 22-bit random component
The random component has two parts:
a node identifier (0 to 20 bits)
a counter (2 to 22 bits)
The node identifier can be provided by the tsid.node system property when bootstrapping the application:
-Dtsid.node="12"
load-module module-echo-cancel aec_args="analog_gain_control=0 digital_gain_control=0"
+
in the file:
/etc/pulse/daemon.conf
or
~/.config/pulse/daemon.conf
Uncomment the line, and set to no:
flat-volumes = no
load-module module-echo-cancel source_name=logitechsource
and then at the bottom of the file add
set-default-source logitechsource
In this case I named the source logitechsource, but you can name it whatever you want and simply either restart pulseaudio.
Caused by: java.lang.NullPointerException
at java.util.Hashtable$Entry.setValue (Hashtable.java:1286)
at org.apache.maven.model.interpolation.StringVisitorModelInterpolator$ModelVisitor.visit (StringVisitorModelInterpolator.java:1429)
Seems to be resolved in 3.8.4. I'm now seeing a proper error message:
Resolving expression: '${project.version}': Detected the following recursive expression cycle in 'project.version': [version, revision] @ org.example:foo:${revision}, /experiment/pom.xml -> [Help 2]
‘-XX:MinRAMPercentage’ JVM argument will be used to compute Java heap size only if your overall available memory’s size in the physical server (or in the container) is less than 250MB
The following table shows the values that are used when -XX:+UseContainerSupport is set:
Less than 1 GB 50% <size>
1 GB - 2 GB <size> - 512 MB
Greater than 2 GB 75% <size>
The default heap size is capped at 25 GB
The default heap size for containers takes affect only when the following conditions are met:
The application is running in a container environment.
The memory limit for the container is set.
The -XX:+UseContainerSupport option is set, which is the default behavior.
We observed that generally the embedding representation is very rich and information dense. For example, reducing the dimensionality of the inputs using SVD or PCA, even by 10%, generally results in worse downstream performance on specific tasks.
moving:
to the end of the command: ctrl-e
to the begin of the command: ctrl-a
forward a word: alt-f
backword a word: alt-b
deleting:
from current cursor position to the end of word: ald-d
from current cursor position to the begin of word: clt-w
If you upgrade your spring-boot version to >= 2.3.0 you can enable the liveness and readiness probes by adding:
management:
health:
probes:
enabled: true
You need to manually import the ServerHttpSecurity invoke.
import org.springframework.security.config.web.server.invoke
/
import org.springframework.security.config.annotation.web.invoke
A Dockerfile that performs a Jib-like build is shown below:
# Jib uses Adoptium Eclipse Temurin (formerly AdoptOpenJDK).
FROM eclipse-temurin:11-jre
# Multiple copy statements are used to break the app into layers,
# allowing for faster rebuilds after small changes
COPY dependencyJars /app/libs
COPY snapshotDependencyJars /app/libs
COPY projectDependencyJars /app/libs
COPY resources /app/resources
COPY classFiles /app/classes
# Jib's extra directory ("src/main/jib" by default) is used to add extra, non-classpath files
COPY src/main/jib /
# Jib's default entrypoint when container.entrypoint is not set
ENTRYPOINT ["java", jib.container.jvmFlags, "-cp", "/app/resources:/app/classes:/app/libs/*", jib.container.mainClass]
CMD [jib.container.args]
v1.6.0 has been released with this feature; prefix your source image configuration with docker:// to use a base image stored in the Docker daemon.
Gradle: jib.from.image = 'docker://docker-image'
Maven: <from><image>docker://docker-image</image></from>
Avoid Pinning to latest
You shouldn’t consume the latest tag of an image whenever a more specific alternative is available. ...
If the author doesn’t maintain latest, you could end up with a heavily outdated image ... Several container ecosystem projects now warn against using latest for this reason.
You want to be using argon2id.
A KDF is a function that takes some input (in this case the user's password) and generates a key.
Good KDFs reduce this risk by being what's technically referred to as "expensive". Rather than performing one simple calculation to turn a password into a key, they perform a lot of calculations.
However, there's another axis of expense that can be considered - memory. If the KDF algorithm requires a significant amount of RAM, the degree to which it can be performed in parallel on a GPU is massively reduced.
/store/book/author $.store.book[*].author the authors of all books in the store
//author $..author all authors
/store/* $.store.* all things in store, which are some books and a red bicycle.
/store//price $.store..price the price of everything in the store.
//book[3] $..book[2] the third book
//book[last()] $..book[(@.length-1)]
$..book[-1:] the last book in order.
//book[position()<3] $..book[0,1]
$..book[:2] the first two books
//book[isbn] $..book[?(@.isbn)] filter all books with isbn number
//book[price<10] $..book[?(@.price<10)] filter all books cheapier than 10
//* $..* all Elements in XML document. All members of JSON structure.
The purgeServerSideCache method is deprecated and calling it has no effect - you'll get a console warning about that. This method is now replaced with refreshServerSideStore