Skip to content
Toggle navigation
P
Projects
G
Groups
S
Snippets
Help
las3_pub
/
predictable_parallel_patterns
This project
Loading...
Sign in
Toggle navigation
Go to a project
Project
Repository
Issues
0
Merge Requests
0
Pipelines
Wiki
Members
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Commit
779978e2
authored
Jun 18, 2019
by
FritzFlorian
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
Fix Bug in scan (elements where skipped).
parent
73550b12
Show whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
57 additions
and
31 deletions
+57
-31
lib/pls/include/pls/algorithms/for_each_impl.h
+1
-1
lib/pls/include/pls/algorithms/scan_impl.h
+56
-30
No files found.
lib/pls/include/pls/algorithms/for_each_impl.h
View file @
779978e2
...
...
@@ -15,7 +15,7 @@ namespace internal {
template
<
typename
RandomIt
,
typename
Function
>
void
for_each
(
RandomIt
first
,
RandomIt
last
,
const
Function
&
function
)
{
using
namespace
::
pls
::
internal
::
scheduling
;
constexpr
long
min_elements
=
4
;
constexpr
long
min_elements
=
1
;
// TODO: tune this value/allow for execution strategies
long
num_elements
=
std
::
distance
(
first
,
last
);
if
(
num_elements
<=
min_elements
)
{
...
...
lib/pls/include/pls/algorithms/scan_impl.h
View file @
779978e2
...
...
@@ -7,10 +7,14 @@
#include "pls/pls.h"
#include "pls/internal/scheduling/thread_state.h"
#include "pls/internal/scheduling/task.h"
namespace
pls
{
namespace
algorithm
{
namespace
internal
{
using
namespace
pls
::
internal
::
scheduling
;
template
<
typename
InIter
,
typename
OutIter
,
typename
BinaryOp
,
typename
Type
>
void
serial_scan
(
InIter
input_start
,
const
InIter
input_end
,
OutIter
output
,
BinaryOp
op
,
Type
neutral_element
)
{
auto
current_input
=
input_start
;
...
...
@@ -25,52 +29,74 @@ void serial_scan(InIter input_start, const InIter input_end, OutIter output, Bin
}
}
}
template
<
typename
InIter
,
typename
OutIter
,
typename
BinaryOp
,
typename
Type
>
void
scan
(
InIter
in_start
,
const
InIter
in_end
,
OutIter
out
,
BinaryOp
op
,
Type
neutral_elem
)
{
constexpr
auto
chunks_per_thread
=
4
;
using
namespace
pls
::
internal
::
scheduling
;
// TODO: This must be dynamic to make sense, as it has a far bigger influence than any other cutoff.
// The current strategy is static partitioning, and suboptimal in inballanced workloads.
auto
size
=
std
::
distance
(
in_start
,
in_end
);
class
scan_task
:
public
pls
::
internal
::
scheduling
::
task
{
const
InIter
in_start_
;
const
InIter
in_end_
;
const
OutIter
out_
;
const
BinaryOp
op_
;
const
Type
neutral_elem_
;
long
size_
,
chunks_
;
long
items_per_chunk_
;
Type
*
chunk_sums_
;
public
:
scan_task
(
const
InIter
in_start
,
const
InIter
in_end
,
const
OutIter
out
,
const
BinaryOp
op
,
const
Type
neutral_elem
)
:
in_start_
{
in_start
},
in_end_
{
in_end
},
out_
{
out
},
op_
{
op
},
neutral_elem_
{
neutral_elem
}
{
constexpr
auto
chunks_per_thread
=
1
;
size_
=
std
::
distance
(
in_start
,
in_end
);
auto
num_threads
=
thread_state
::
get
()
->
scheduler_
->
num_threads
();
auto
chunks
=
num_threads
*
chunks_per_thread
;
auto
items_per_chunk
=
std
::
max
(
1l
,
size
/
chunks
)
;
chunks_
=
num_threads
*
chunks_per_thread
;
items_per_chunk_
=
size_
/
chunks_
+
1
;
scheduler
::
allocate_on_stack
(
sizeof
(
Type
)
*
(
chunks
),
[
&
](
void
*
memory
)
{
Type
*
chunk_sums
=
reinterpret_cast
<
Type
*>
(
memory
)
;
chunk_sums_
=
reinterpret_cast
<
Type
*>
(
allocate_memory
(
sizeof
(
Type
)
*
chunks_
));
}
;
void
execute_internal
()
override
{
// First Pass = calculate each chunks individual prefix sum
for_each_range
(
0
,
chunks
,
[
&
](
int
i
)
{
auto
chunk_start
=
in_start
+
items_per_chunk
*
i
;
auto
chunk_end
=
std
::
min
(
in_end
,
chunk_start
+
items_per_chunk
);
auto
chunk_output
=
out
+
items_per_chunk
*
i
;
internal
::
serial_scan
(
chunk_start
,
chunk_end
,
chunk_output
,
op
,
neutral_elem
);
chunk_sums
[
i
]
=
*
(
out
+
std
::
distance
(
chunk_start
,
chunk_end
)
-
1
);
for_each_range
(
0
,
chunks_
,
[
&
](
int
i
)
{
auto
chunk_start
=
in_start_
+
items_per_chunk_
*
i
;
auto
chunk_end
=
std
::
min
(
in_end_
,
chunk_start
+
items_per_chunk_
);
auto
chunk_size
=
std
::
distance
(
chunk_start
,
chunk_end
);
auto
chunk_output
=
out_
+
items_per_chunk_
*
i
;
internal
::
serial_scan
(
chunk_start
,
chunk_end
,
chunk_output
,
op_
,
neutral_elem_
);
auto
last_chunk_value
=
*
(
chunk_output
+
chunk_size
-
1
);
chunk_sums_
[
i
]
=
last_chunk_value
;
});
// Calculate prefix sums of each chunks sum
// (effectively the prefix sum at the end of each chunk, then used to correct the following chunk).
internal
::
serial_scan
(
chunk_sums
,
chunk_sums
+
chunks
,
chunk_sums
,
std
::
plus
<
int
>
(),
0
);
internal
::
serial_scan
(
chunk_sums
_
,
chunk_sums_
+
chunks_
,
chunk_sums_
,
op_
,
neutral_elem_
);
// Second Pass = Use results from first pass to correct each chunks sum
auto
output_start
=
out
;
auto
output_end
=
out
+
size
;
for_each_range
(
1
,
chunks
,
[
&
](
int
i
)
{
auto
chunk_start
=
output_start
+
items_per_chunk
*
i
;
auto
chunk_end
=
std
::
min
(
output_end
,
chunk_start
+
items_per_chunk
);
auto
output_start
=
out
_
;
auto
output_end
=
out
_
+
size_
;
for_each_range
(
1
,
chunks
_
,
[
&
](
int
i
)
{
auto
chunk_start
=
output_start
+
items_per_chunk
_
*
i
;
auto
chunk_end
=
std
::
min
(
output_end
,
chunk_start
+
items_per_chunk
_
);
for
(;
chunk_start
!=
chunk_end
;
chunk_start
++
)
{
*
chunk_start
=
op
(
*
chunk_start
,
chunk_sums
[
i
-
1
]);
*
chunk_start
=
op
_
(
*
chunk_start
,
chunk_sums_
[
i
-
1
]);
}
});
});
}
};
}
template
<
typename
InIter
,
typename
OutIter
,
typename
BinaryOp
,
typename
Type
>
void
scan
(
InIter
in_start
,
const
InIter
in_end
,
OutIter
out
,
BinaryOp
op
,
Type
neutral_elem
)
{
using
namespace
pls
::
internal
::
scheduling
;
// End this work section by cleaning up stack and tasks
scheduler
::
wait_for_all
(
);
using
scan_task_type
=
internal
::
scan_task
<
InIter
,
OutIter
,
BinaryOp
,
Type
>
;
scheduler
::
spawn_child_and_wait
<
scan_task_type
>
(
in_start
,
in_end
,
out
,
op
,
neutral_elem
);
}
}
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment