- 28 Feb, 2017 1 commit
-
-
Michael Kozono authored
-
- 03 Jun, 2016 2 commits
-
-
James Lopez authored
This reverts commit 3e991230.
-
James Lopez authored
# Conflicts: # app/models/project.rb
-
- 23 May, 2016 1 commit
-
-
Robert Speicher authored
-
- 04 Feb, 2016 1 commit
-
-
Josh Frye authored
-
- 27 Jan, 2016 1 commit
-
-
Yorick Peterse authored
By simply loading the first event from the already sorted set we save ourselves extra (slow) queries just to get the latest update timestamp. This removes the need for Event.latest_update_time and significantly reduces the time needed to build an Atom feed. Fixes gitlab-org/gitlab-ce#12415
-
- 08 Jan, 2016 2 commits
-
-
Robert Speicher authored
-
Robert Speicher authored
-
- 18 Nov, 2015 1 commit
-
-
Yorick Peterse authored
Instead of using MAX(events.updated_at) we can simply sort the events in descending order by the "id" column and grab the first row. In other words, instead of this: SELECT max(events.updated_at) AS max_id FROM events LEFT OUTER JOIN projects ON projects.id = events.project_id LEFT OUTER JOIN namespaces ON namespaces.id = projects.namespace_id WHERE events.author_id IS NOT NULL AND events.project_id IN (13083); we can use this: SELECT events.updated_at AS max_id FROM events LEFT OUTER JOIN projects ON projects.id = events.project_id LEFT OUTER JOIN namespaces ON namespaces.id = projects.namespace_id WHERE events.author_id IS NOT NULL AND events.project_id IN (13083) ORDER BY events.id DESC LIMIT 1; This has the benefit that on PostgreSQL a backwards index scan can be used, which due to the "LIMIT 1" will at most process only a single row. This in turn greatly speeds up the process of grabbing the latest update time. This can be confirmed by looking at the query plans. The first query produces the following plan: Aggregate (cost=43779.84..43779.85 rows=1 width=12) (actual time=2142.462..2142.462 rows=1 loops=1) -> Index Scan using index_events_on_project_id on events (cost=0.43..43704.69 rows=30060 width=12) (actual time=0.033..2138.086 rows=32769 loops=1) Index Cond: (project_id = 13083) Filter: (author_id IS NOT NULL) Planning time: 1.248 ms Execution time: 2142.548 ms The second query in turn produces the following plan: Limit (cost=0.43..41.65 rows=1 width=16) (actual time=1.394..1.394 rows=1 loops=1) -> Index Scan Backward using events_pkey on events (cost=0.43..1238907.96 rows=30060 width=16) (actual time=1.394..1.394 rows=1 loops=1) Filter: ((author_id IS NOT NULL) AND (project_id = 13083)) Rows Removed by Filter: 2104 Planning time: 0.166 ms Execution time: 1.408 ms According to the above plans the 2nd query is around 1500 times faster. However, re-running the first query produces timings of around 80 ms, making the 2nd query "only" around 55 times faster.
-
- 21 May, 2015 1 commit
-
-
Douwe Maan authored
-
- 23 Apr, 2015 3 commits
-
-
Douwe Maan authored
-
Douwe Maan authored
-
Douwe Maan authored
-
- 31 Dec, 2014 1 commit
-
-
Jeremy Maziarz authored
-
- 04 Dec, 2014 1 commit
-
-
Dmitriy Zaporozhets authored
Signed-off-by: Dmitriy Zaporozhets <dmitriy.zaporozhets@gmail.com>
-
- 09 Oct, 2013 1 commit
-
-
Steven Thonus authored
-
- 05 Apr, 2013 1 commit
-
-
Dmitriy Zaporozhets authored
-
- 10 Mar, 2013 1 commit
-
-
Dmitriy Zaporozhets authored
-
- 27 Jan, 2013 1 commit
-
-
Dmitriy Zaporozhets authored
-
- 14 Dec, 2012 1 commit
-
-
Dmitriy Zaporozhets authored
-
- 24 Sep, 2012 1 commit
-
-
Alex Leutgöb authored
- Add project name to event title - Push: Entry links to single commit or commits overview depending on number of pushed commits - Push: Display first 15 commits with commit message and author and link to single commit - Issues: Display issue description
-
- 05 Jul, 2012 1 commit
-
-
randx authored
-
- 12 Jun, 2012 1 commit
-
-
randx authored
* dashboard should be in dashboard controller not project index * projects index removed
-
- 01 Jun, 2012 1 commit
-
-
Nihad Abbasov authored
-